var/home/core/zuul-output/0000755000175000017500000000000015135706544014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015135712031015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000211304115135711755020262 0ustar corecorewikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD  ?Eڤ펯_ˎ6Ϸ7+%f?長ox[o8W5օ!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPFS]dP>Li.`|!>ڌj+ACl21E^#QDuxGvZ4c$)9ӋrYWoxCNQWs]8M%3KpNGIrND}2SRCK.(^$0^@hH9%!40Jm>*Kdg?y7|&#)3+o,2s%R>!%*XC7Ln* wCƕH#FLzsѹ Xߛk׹1{,wŻ4v+(n^RϚOGO;5p Cj·1z_j( ,"z-Ee}t(QCuˠMkmi+2z5iݸ6C~z+_Ex$\}*9h>t m2m`QɢJ[a|$ᑨj:D+ʎ; 9Gacm_jY-y`)͐o΁GWo(C U ?}aK+d&?>Y;ufʕ"uZ0EyT0: =XVy#iEW&q]#v0nFNV-9JrdK\D2s&[#bE(mV9ىN囋{V5e1߯F1>9r;:J_T{*T\hVQxi0LZD T{ /WHc&)_`i=į`PÝr JovJw`纪}PSSii4wT (Dnm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁weor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/P_]F@?qr7@sON_}ۿ릶ytoyמseQv^sP3.sP1'Ns}d_ս=f1Jid % Jwe`40^|ǜd]z dJR-Дxq4lZ,Z[|e 'Ƙ$b2JOh k[b>¾h[;:>OM=y)֖[Sm5*_?$cjf `~ߛUIOvl/.4`P{d056 %w ^?sʫ"nK)D}O >%9r}1j#e[tRQ9*ء !ǨLJ- upƜ/4cY\[|Xs;ɾ7-<S1wg y &SL9qk;NP> ,wդjtah-j:_[;4Wg_0K>є0vNۈ/ze={< 1;/STcD,ڙ`[3XPo0TXx ZYޏ=S-ܑ2ƹڞ7կZ8m1`qAewQT*:ÊxtŨ!u}$K6tem@t):êtx: `)L`m GƂ%k1羨(zv:U!2`cV, lNdV5m$/KFS#0gLwNO6¨h}'XvوPkWn}/7d*1q* c0.$\+XND]P*84[߷Q뽃J޸8iD WPC49 *#LC ءzCwS%'m'3ܚ|otoʉ!9:PZ"ρ5M^kVځIX%G^{;+Fi7Z(ZN~;MM/u2}ݼPݫedKAd#[ BeMP6" YǨ 0vyv?7R F"}8&q]ows!Z!C4g*8n]rMQ ;N>Sr??Ӽ]\+hSQזL c̖F4BJ2ᮚ苮p(r%Q 6<$(Ӣ(RvA A-^dX?+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDAWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44JٵK+Woc.F3 %N%FF"HH"\$ۤ_5UWd̡bh塘ZRI&{3TUFp/:4TƳ5[۲yzz+ 4D.Ճ`!TnPFp':.4dMFN=/5ܙz,4kA<:z7y0^} "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓoJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A?7 /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"V[^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{7+ߑ`xV,)ޖ,3~JPͪm|$oV1yU<̐t6 T m^ [IgINJ\Оf*Z"I)+>n#y 9D*A$$"^)dVQ.(rO6߿Yw_Ȣaޒu'->Xmw,*=.[G n >X9Ī;x7%dT:`ٓ~:VD)O>UD;;]Y,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkW%!vĉq|?mtB|A ?dXu7LGml?*uTC̶V`FVY>EC}DnG+UaKtȃbeb筃kݴO~f^o⊈ 8MK:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȟ.vIH"ŐR ;@_AH<%Ҝ ܣTvnVUY om?'4%hs.o&˛Sy*LD ZmWb{ݏa ې!rGHw@56DǑq LA!&mYJ*mxz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1]u6d逖`7zGMf}k/⨼0Κ_pLq7k!dTi XʽAQGwG% C<ˉvRO)?J=y:1V(!L7,RPEd;)QϢ +RlWDžuF7LFֆoM~ar*EtIbW>jqour?qzJJaQ#-n`/$fhnqgTĔO5 ꐌSYXzv9[ezksA`<dkON৯s|&*pNaJه5B5H:W2% `6MRR'xZtfC$1aH_dx$1'/v^ZZ4`9);q`F"d1v>ժbLGd~MP%m x52LMF9 E"A,S Vo}\"X.2< 5FB΢u.`aJ#Tk’"D#cuCXȉ4 ՖK(KP|dZ1&8{9rLnMRф%V Ng2K|`ot.GSGd oE'!B'Nb1{8LW^9KbN;sö!`0ݘ/l+1L#B8U֕&*?V6N{դ}Y(INBKhx2 *MOenT.a~.E jG)j{=u^K+Ȫcv/w#MivX :)ǪCZUnAS`SK6OSxa3 W; K>窜̀'n 3u0?K@BS %fee}i]>̤+*l:\歶!IZ5>H;0)N.w7ߍ|+qUߤ^oå~4en\.cY[s'wSSۘf ?.D s}Y~/J[}jX^ޗ_-/̍ݥ*n./cus}]\>\\^'W_nAqC_oO-S_sOq?B}mmK2/@DJt}=xL@5MG0ZY,\S Eb uw:YɊ|ZԘ8'ˠ*>q/E b\ R%.aS qY>W Rlz!>Z.|<VD h5^6eM>y̆@ x>Lh!*<-lo_V684A飑i2#@+j3l૎S1@:G|gRcƈ?H(m>LC,HI~'.Op% ' c*Dp*cj|>z G` |]e*:nq!`{ qBAgPSO}E`́JPu#]' 3N+;fwt[wL X1!;W$*죓Ha-s>Vzk[~S_vD.yΕ`h9U|A܌ЃECTC Tnpצho!=V qy)U cigs^>sgv"4N9W_iI NRCǔd X1Lb.u@`X]nl}!:ViI[/SE un޷(ȊD0M^`MDN74Т C>F-}$A:XBgJWq&4ۓflq6TX)ى?Nwg>]dt*?Ű~{N_w7p682~ =WBX"XA:#u-9`x 92$4_>9WvTIj`+C2"s%DƖ|2H\2+AaTaBˮ}L@dr_Wfc>IdA Od[jlec=XJ|&+-T1m8NP$%s,ig\Z:h Ћ߉n!r}_\ \5 6 d#=&X^-kOwĝJO\Vj; )!eoB4F\jtctUb.L[3M8V|&jZz/@7aV),A[5TpUZL_?CU0E [%W%vl x٘3܎y,< )i7 Ո: tC`\?c%v7\Ct!$9iç$><+c~݊lz1H[E'2/clQ.I`AWOlw&5fH n`gMytdx)lwAK~GgbJI-tq5/i ?WǠr^C/1NEU<=co(k0Q~wˌ\g,\ rf\PUH,L#L7E"`0dq@zn~+CX|,l_B'9Dcuu|~z+G q|-bb^HcUha9ce1P[;qsA.Ǎ-]W‹y?ڕ^Pm:>I+Ȧ6' ,}U=̀*Eg.6_~OJ/8V ?ç&+|t><,BLqL򱷬dS{X6"X#-^䀕#{К4i̎'QIc(<ǩJi lc*n;YKOIXA|i޵+ȗ%Qntil8E@_f$ٱӤt ]]>eCu׬ 45mp0 C'x$ӔZRku I2$g }98ǦnڗyiZs3,׎K4f8,^E1LpzӮLd2@Ƿs܃@IVy2'PGv6?f'2Կ6MX$ ?.yrJvX F(c{ηu()hPg-ְ,^%OLWw? *Ayy>˳xZino2@\gsq!æOb"ry_ԥil-[f$Q\6 l{tdd,b97 x(q4eHSI,8>ZΥַޮI' O.jztlNDtIky "FzDsl"),}{g!y͖ºM 9,vID'A"oO-.?W? M& UE! 4z O>08kKyt ?OOI0PjmW65 )+LwԾPIݶ05 &])W"] نÓԉԽMaNxRVyKfRjuɳ*y'g,{S*ώbFYioZ`"V!UY-&VZY^ʥT?L{:G4ª>q Bbu eK^-U W4(Q=LW[ 4iQ)42C$ [wwi䕔zv1М'd1}5µ4KLA)mm͞т8w5JcFvdPyF?ɽ#=ڑowxgl;|gtLek'w =s `zS]͙3PGQH:@w,-0p On3Z.ƠM0Bvee{ÄaVD'IPs s=OD<'x=#""mϻ]6Z*edKo e_3u4D.4e϶˚|I,8>ZnJ4?.sP?]]5= uԙKj =\Vٻ޶q-Ww7E)wIL&"%9Զ49$%ӎڡڨtj[y u<WR +OٟῖYuʇ;{Qe ')7?xl{̇@]u4+Y/O(CCe#1dokO+qC{D%Tǰ>Tb 룁'}z\zvxãr[В@]QQ#nm[i ]tKA@^4ˈ'fphtʱu[u0Nw'3 ;@ +>Feߕ8 ڣϧ潮8 C`j{q]3H9xpZ;ED ,t.Hl4YǸ!K:]C'B tHnxhݰ#Mu2 v4X-7rpe +D,b`Ѵz粞5E"pPg 6}NK&@^2&z hJ 10A:P4v"+u% 6rtN V y80 ȌPQՕpΪ^^pV,pkHڲ%l"˔|Y^6mM6mXŽӂ?41eϤcEsQ`BH-}&qZ䄐5VOGy2Q\Jq碍)q2š%GS 6@4aJ@_lCQIIrƽkL/p|2W wrs|}tx /1I˲,lVcA P|)\dAL`4;u~ "3p2#)?9ҁIּSKrıo`Twp:hl{@th&+tՓI=5Nql&Љ}30&~f{AOMr7g<ӲUi 0rYƆrY0tHЄBrҡeyEj+Pd ד~Ys4/edI>O ٨"C8pE`d}! 5QVc{ȩ>ո&A-n]X`t ĸۈ *>=*urV*pm[]XicrRcM* MA"H#,1;j5wfl"?iϵc.KN/~ϦiQ ?@p$'/a_O*X)^:KlS8ڰ7I6);Y$p0ɥv/3qsÒOr#ytTD:~*g9+XW]\ehwO5rqK]~lZ4+L\wDal8~$YBι @~rLk8i{UDt,Y+u| j0c/L7RCY& O;S R"CR'rt_dixD,ulpv=oCX׸W4ywٌ8VRf[[iV%Kj[JSg3F!E~aBq@`Ĺz@}(g+Qβ1Cy)Zwk)q\eL,gβAlinE2niM!Z&u~DާxeUS,y)g1u7t|8OthZLZ71l K30Mb"gV2CQd,uQX*Բ)dE6uea9pviդƢl*Xۋ`,+pYM#/^+ jit%qp{i{ci؜0? =+C;h0ah77S@dRG^ !X}UT) vRKHw0ڎf[Y[Y-y ~Zp^ڱ_ht olQh`3IT.qoӣKSV9]LH INI\/?aou tPtwB{Jw!tK.}JT^v+ݩ_;vݧkڷ{w.kRk@;HށPgPgB uF';N4B= uw [%ہPowBI*?POB V v 4؝i{@hJhO#4ܓpB*lB$m tI?6Ul Piq;%?ͭ^:QC碾cMw9-.<MV7yUh9bt bN }'Jxc0d7" ˬhFjRz:={^C_ oh)׿u{Aϳzv/K9/}q>wUtDfe^eM>]|Gwĥ!@=ٲ{E^0( qs(}p N|u.|̭-N("CU20Pfs?IUi衣aqa>-&uԦPnޯ>Z> ,!*H5S goUmt:K??jawh[YH1+KdC?[ΘAE>7Ɓ!܎8I^5Γ@*%xKxL[yhyݻs_Fd2="e ڲ"[nF@T[9Zwg@Z΁ybFb }, 2nLE+KZ[_::/L2銇 "|Ymֈx9, ͮIVc`T-dzNr7ኘ4^BʆKfu]? h s\S̰cK&-b\PK0(X=˃_q"$e8'З@k$> "@ OBvYx(I =4E@w P+'3 ~f\bg3tiL`ᡱA)hyޭAz+ =`MӉGEz՛|ds-7ˍ` Gι_]gkL, s֑l\RN.p#ELF0۲:bJ%̸} L Sn]YKfz{U4׬:XCK S\*]}I1pi`a6A wQSTs-#Osta蠘OMg w,jW^8Bٮ8Bf(k Z!5kC639Iu7MRB8ݩ@y' +|Y0b͡D>B>-@Lao8stlw#SlwɔCEt55`ss>`aa@ԬYX:w|OVHNzUs\a[Gh(o~)aڷ:!VȦ)2ܥZW+ϬJ<\Hb7B?) y[ӭD,5ymM춹Se7c*UjxgCXL}jMԳWK_{as7K=Ƈ^<I}E>#5g-hk.'s!@vpفs!0xBv_ 0!F6'q~JL抪^gS8/g1w- t9Cv֛T}icWc+aCG5e\Īrq2K/A]iT-hCm/>X:evj?a. W帆0GO+}d8kv:X-2ͧ|6+cydEY!p`>W rT \5)wܗE|U [m+XrB$9~61[٪}/mد7P;߳l]kw3K?>x"Wrӯ$]r|]w?ܣv)i4SsQs]aZC^ܫ&_$re3 %=ɮT8yC210~qKL5#$l8xmE`CM`%U4WOo0g*Y6ʠs|F9hEk0R?ჳABi<ݸb|pڮB;>8D)Q1Sɧ0e*Gp,uPQQ'8xq=oATNVm̳*:8 S~,óxG\V4j{=fx(THTMKɱWbTP=&?N@YM g m ʓL1B̋.BE)GX Sp({N$J,G޵-$_\^ 2bsJE:~/  ELVf\1/א֋?ژ%[`[cmCH㡎ay^q< !0_&j5(c^4!)Yw=&_~=B8k;:-wc 1i#ojW\u]i^poŃ/Ipb=ğ+ O{-8.3p 5 F1YWkzqGÑ"  ZOQĄWN} ̓@&-e}JDI0xN)Z(A"輖OD;l;a}mMtv[v 18k[:(&"=ˁUsmD߲`m`ðct)AI+(b^X)|5Zy_nU3h | )0^΂WoN-_bܵÆY$䃩1BC7 ݑdtBҺƩn!Alҷ-qBdɯ) 9_ TGp*nd] ǵ|HoadeP6&"5uNA$<=|"SUBSQO9ŋLkbHN'J^ Uh$(~ i<'׭Ȣ^ch%Q|vGtv㶲x=:egŃMźqǣ%c;s6RgTM;-;ٽJ?9_fˢ NB6R@ B5Ee9\Avae(ڏݸKciO0 s[I  YxF/lRCϹML"E rnjݚs rh^ t X5(b8 z\F7!эV 6Htn`̃U,8O_—N7DҤڑ }s# qWiKUrþ8yҎwM'C˫{yx|)$BNXG` O'VڅHi=C}/,8N9: =A;+]("+MɉQļMpu>9曓1f1SBhBJB)Em1e `[CTdG(M1>qeaarGB1*Ё^\ b"wX q';4.-!~-w 4g "By&H:O!; g.9UT($9=p܌;s"O[`ɁT [ug# ^x"ʣy3r!7>TYEH-cr2,=" E) So7;6OPIPZ *p!j0T٦f-E w/SF_H1sād:,UmZg!kR!4}G%zssL953R TkLL R^xtEyNN:Uq.EҌU\ ˀh񩛁':=[FKi @C[$XCO&y?,N.L łl/‡xJ"%1.g Wqu*kQ$3*`Yp b2xz 5g<:D|{!SN &f6\K:kM6=X`Qxl,ђP$!y3S=jyТt'ZRdJ#j$6$7N<= V_GGҙ# ͘ BVRx)y~a1ȫ GqMTWeEyhW-Wː?^d1lUײh][0~ aW@9C$֥#^0&f?.,h-Ulf_!ҡM-δoN8{^IkJ%j RJ6;HÝF"^v'Ũr녭n =HiuxʂWWN}O 2|. нY0 ,$JŠ0J 6%o6{|ȖJ̞Y ǣ_ I lt6|%1>%m0ᕾ.ErBzw Xp49;ήy1j毹s8o%)Zv?vvyO=S w_j& &;,&w/'On.ڡn}|*juSV뎡^FG;&߭$Z'|{6t_uBOw(~yJ/U>|a?ٍ.Ȃ"hMU#ƺ* 1HzU/2fi}?db北7:UCn ?z.f˥LurpJͻʦ}j1|NMgGWql OzVc%} -Y43)=gk7̐21d|,8Nk&R]qGarZ;l$=SY /mZpCRC@R@6BV 1^%WfNY?հ8y!ka46Ƚ "5uN$ͅv>$ֿFaѦk'^= QYV2iN^7v _P=%( T9 Bܻ(up+j' đb`];Cn<&5CC$KD?s4dXx( QuTZԪ/&&b"|wLβ3D}Da+u}*neA3k¹Um= m_.$/vD1@d1b a cdH&6Ih?P$(<"EIefJ)sWE(~H"AZ1ʂ~Y;`Ek%Ci2:cw{z|brfNO*ҽ ?w:& tEX ́l$M )u鋤yP.?c\)rreW&5eԤBsT PCJJ( "T=OZid^-.sI(wU1;0K5*- i}S-\t읪0ϛ9v]99nrtb %tXKS CIye\K"𣰤:Ƃ41꼼R&Yp]8`89xC zՆ%R\8~Xy;W̿LŽy %*mY8ꫂZ[+U@oGbF:ߍu1px=Co}Kǂ 7]ZyY:Kۗʭ}@7cֳ8t"T (yIE%E3[!R\$?.U%̈TJd\xUv? @K5ZLz _F>(o9>xӌMޟzU5dl7b\=r9!gOkPp|f6GCՏGO/eȮyc$uǙ^mc4}tzp5= _L -ƣ>i%"? r9 ǝ#oq!hAm0֎` cp~ z v>ꃹ pLt t{~ƒIsA' B$&T1"mEAT_oF.z#=>`jH:o[ q<ְDX/o5@DUpp q&@VV wUcB262LCl"Q".4/he,cI9j=W7\&Y =H6 4P`K`ta 僉Ft,Csxmr41j6{6T eWA-_n&Wnr΄t'P {o4 /͗a}^NZ*8}pQzn4>ʍ*h:I~-Jߖ/5q':xEukz+3ߎ-z#5l(sA1rkH@ S+_W$JJvn>Ivy+~3y(9\%ZW"‡wF:<%,Ԋ^;uсQeM҉ a8([eۆM<1;hbCc*ۤ.#&@* #<1+3<'r5 wP?9mrۏؿ@nкD* w1j͐)Ey&5@XIv&@t*Lm106&ZѰI|[-?%o;&R솇͆wPHCsWoy{aF;0-17U޳2O果iS`HX|ac׃[-O'<tKy^-]ޖx~ɶN{hZ;pR{p[EP߹/r[2g_,f,fb\[ fċؖRDZ.+M6mk4ThF7查sN%+,w N<? $i-@.v3-\*{VؽHQX?[n!VZ\ :G |D Sd␘!)AbG$n'0<8WS]'b,P5$ȫ?V:g^_?_ ]%G Tn?[t4S{%.ˑUQ2*JT eL )oS*X:HjP_·rnD.>w#x:vg<3ļ9oA@DkL&.4dh-HƲtG,&LCh12OR)Vc,[ !A1ARʜbYqisWՒ?X2QZ%cY1Tg%"q)N!( PaeuZ1O,u;%u;!D}HƒR,2 2’ T U2&%FLaKHsnkF.idddp!HƲ,ׂdb9;D^k_OAlR3kSDdWՓ?X24b=B[N$Uu}u18Szn4> b$P|u"op'{:K_|k߹㞛?W[1GYvW cT#fp $*$}.!Bk;EzGZhހ󄴄6( rݔ6P 2$陫>$17]`p}`1 .7(tT3bįC7&8i 1؂X@ N?w~?_? -픿޺܍\?sUmof/ ;@ۡIEϤs-:3gje,2xzV Jɠ7Nu?h<%;{N&\x[K#@6\rkf@&î郹ʂWt ߈oj!9Epu~x̗֊.MQ*.ƲQ6 N1wr$NQ!S^Tdl6nah"F+*뼷/^D n#D"/uX6XOT7sVYkԻ?_\OD>+^/O3Xj`ژTؔRIs͸skrH9VŒ-9윜xy}kd&O uy 3z/l9V]\M7V]ZN꨿O k@S;c3 \qQS&[{$A9 YlHq&YCKgUJm<&M|Bq 'P|ZDz[Z˵B)"ifSy#F)$u-j -[T[C u 6 65O:CDRG +5e4YʩiS)c*tDEUyZËk͟FJ1L) ! :,S2 gq)D@I(I­cNQ 8\H#j!ZlN֤"7Z:REw&7ËΡ\¸Dίdv&JҚ r92bsDnƝQӋCmkt]AMrq`٠ ռ1.Z%} @Z?kʚ"L" odJ '/Ԟ "/(gA(6H%6G%{rw}J-cu%DRPKj)V/<ڽ%_Зn5$[!^V=HvuKK%'xk@&ׇc x!6-0JW{3&ZMV1E"G0u"w(rIr _\t/#u?ymҔ'P?{۸俊`%ALf%sL# 6ٜэ-Dyg0(ɦ٦%#=(WUB arafYHyf20^y`J9} (݀DTJw cow$\n:RZ#e3A >Щ`z*Z,Y. >(LaPPp͚\sEUA0&E\+ךXZ9(gZx FƁn7ƚKPCx"$(5n0Tz$:<+ Q zS-y$G҇ԃL=bVOhzD|ukL|Dw_ P)1xn pҌ! h{`F.:7-Zi@k; j>8U1h!^)Ta% xFv1:h?!, o`ޘ,+8V5Rt6cs\F6̨<)74CrHy*j Ҽ\֞qk(l( }JgNS5NL'GXFw뉄gBu 6nԝuH܀% 6eES3LxlG*&d`vuaBm#vR}(_]G+ERdTJfJW'|W܊qU6y`J2qBHTGx5DTV4>  | ㉄@$\:+PX%R\- Ss(}fxAS|Z9rG)Okwr'IiPY̬pGK RSxXfI1`Oo>)4_ʉFcAY*Z9 eN٣ၥHK- n 7ߣ L@0-RSv8[$*j\fvszH`{ YWpɆC&'iΠ^?1FlL&R$M5tӯ949fuv8~}P^݂ba#&u&pDQֈibC 7 5sڷ(VE}O[52E_}2H;qI`!osgLj'׎9105sAAZqpF SO ˵E([*6#C5Sõ]mn\+axHW!p%@xY % |ʝᅰVY;-,BʂI:1wWuBBCBU=8NTʼAh+rk{o/4(2? G;/K98C[!äP`}(3|pL2=WR9{c@iθ0?n-rPʀ$6#7X̀ʮSV1UsR?Kqc:gM5_$Lд5Ф::al( ns ,wj@s2P>yaϙ$6/,8KfZuc '={sOJzj78{ :C۴E7$nc[B7D'kiD&(]*\gAmJnhhjkZiiƍg4]\Td6Q |2?ftY3 tQ8dr8թH-1#XSOwBQCO&|CA 0 C>X%& fdBi0< ֪#&L%MZQUmܮKsqaq7hkfa#;n@8kTZMB06`*#&i>66 005,6Qpq" -w\q(U{IX!N2fzg;H$O8P =JٶC6[I E\0ʴ=@t,>j7&VOuG:^Fs`aޑaC:1 Y kj7H퀸ArݱUӰrTvW>&m@jUtzʊ,d8]hT8`OMLCZ$lN7Xb-Vb% ev]k 6\w}6h@)z]X4HAK4HJ;Sݑ$iEi`u+HЭA /l#툉DI:kePz{CpSxFiDUc몉Ⱥ3 J\oW;*givQtA7_2z_+iaHZ˳,܉ vt⮖p&dޟSwD&8B' Wx)>T^V[Uw,[k<,],bq9WÙC54pеاヷ>Uzƿ!K a ɩ=fXwP'6FXc}*Ng46 D_|Uڗ'A"ZZg#_Dc#hAoa!(4h&|w$$MD=M05on7ҫ:R tKXS'LYث%QL<]چC~}וd#-ߧhݥ|%je T *C"gHl4Lg~1zDAO$vrB ]Hn^Z"ߜT%i"iNm"첚Iڡ%1' ţjS$_!g&J)'T뭿__bW>'a"}U;Q|F_eQOYK_sC!v%(J~}|b-~1PYm݋6$WfV.7wц,#jmח]uˮv?ƒ';EEu|!7?|0rC+ػ'hwfȍ>㮯tHo@FN9c7vz/zx^M'3ӱY1M S@O0q~oE@O+W1F*Pj̷52)j'ìۨm5 Eȷ8tqP[eooɝv#6 tsӦw[Wj5lz]zW.R5vZtՀ5 Ѣ H]gEHuZ 32õ[)TNR{juiZ 8uՂHJk{)m~u^ҷDE"-(r< #{mdb{Tld'o{\'z6;!R%~FdWgT_Ϧ'X|dM(РeOgyBlvzCcR3AlhTQ]kjerO@r?MKX(~ )I5(z9?=d}p*5rINϒM%ЁeL2U4:=O>Q 8.wigD%sCڜpaE=8wߢJJ8# |MZT=~(H\ ZPKi&$rGNr]/MELQv[ (ح^ IH=6I FEƸsbję S2qy4`#4M*TZwyH &pc#Q&fd  w5s~"Hԛ>g{&+j0ƨvqXw8vhNXT=I'\Ҿ(0[EِngN *(d?Gwl!-w(w-# ݭ@$jEOwR0cy0K,-@[); ShJ[QCpD Uu+jTC@XlDqv^ELUخyitPE=O [(YW7J4xHy*e}Ye\^Hy(M.<}%sEJ2CsJ(q;yϦVywȩN>DRfm}=y< /'ksԫ[a8nrqvbFPM0r4/ M+^r)P7E g5{ZRp G]k@΄ Zrf 3ԢLu֪jWe%+N8n3G+Qpt -qxG|*j|LG蕂Z\SQ?~/>in6/Ӭ<\[i?J+25^.Nè5܉ίVaSΫ㏋?m;_Mϖ$q90 `La\d[^aȚͻQ7ƫ<˱T)pw"xx|9=X!stXrM ة t4 p˸y0"؆EOEيZ&VWlSzuz67jYU(tA0=>BI$r0W_IFHdx=_Vl,ͧdŬ\W^v}@uDGWQ68O %Գ#.h,稊/Z 3w%͍ts8XH I_o7D:%T2,~@%rC&ꛟNg6lʿMy}%=D!fK%g8ޙ;~kAg}o{9 v&JEi)얓f7& nbsXs[r!~כMVoRip}8u=_2obXl V])ԜHc$Q:c3zȧJ_jHH+q O$)EAĢV Oxb#'qR5e 0E ȧH5$q cLc$n)` <x/8gcDi-+FT5qq rM@z.%8n=HXmsy_Ot1*H bؽ M1gȧb/urωꇻ bdLFXFQ,8Ԕy5"{(0z[|wxp91>fY~g\SG<=/p.m!1FO?V/p! \cxZT_A;N~uBA[s6f>bo=s\G:񭗙}u˧gZ1< y0Cv<"뺋4>6R/˅^.r֌a>W#([OgXw.ujbWE<_gX-U*l.;j1SAME(7 +NJ#l 622r`bgPSѵ/ܱO޲oi<`򶈮]ȾSxrVD)[r#3I&̱VN˱.;kCH(jc;(hzhǸz/vv[U2ILƒɕ2nf%~ -!d6-qZrC󪹥="ȼ M,U*m--/vW[å[6p]'fZF/:1Sr$!VIaXy#9|"ꆘ2Jq h|1fG p$͔Ǽ}0 ԥ}qno" l'VGa:Bq9_%"7 𜙱F=) Mu#rQVB`ȻRKq|<^hOep LSWnFmHɷ8 |Wr^Zv`lfOi=nSsC7N`5/~TFZ |``TwOVC >k~m0SWsC>.uYi+rw1֠~̤$>l/wN(趗!C6SNt/ePS!!U|CC-H!UU8qK /g/,zݕm1m@oxl]VG\yY3, Trfs@/he A# yJ&RCq=`C̠c胎1O6|*5G PZRתjZT{.&: `lRSQ]΂au"ʰ؅.J\j꣌sqG8E>!/Fs#PX1Z@}/w͌i 2Q-Kd{k x{kȧBg9b kW~:#aNz )bSJyq" yج>~~go}@FN{;vD 3nxly0CԸp ws(|͈)B(ťP ϸ2 QnqRUe=g" y[w{dp F3_7/k#}c. hd>3iᵣ Vd4\U ֛s*\G!de!e EV b9NT#tQE$phJW>k~E>!_-] xQ!n_36zqDc^? ɵ[ݢ@}i<Cm~xѢ]H=ӊ"=(9+j|ՐP% UpգPt<ˈ+娎2^,/@NL[+=>`T 7ֳeO\0|rPs3QV{X]דS~#m-ZJ`lLT+"LŸ]Qg45DwTؖJ k8 - U U 6Df.v(dACg '1rN\!`AOfC[.XG,x–픊ťI\{t?F"3%m,JîWktmwe-cpcl^tn7a;@ËZ\s]w y'bfKU|1uk@Fױ$cbn5ڌB,RRcwP9ڳ,a@r>( QqO eKcy@鐄æ)L"uu"o4̏H|-{"wE(4bR^1a U_u1: G= BPS! ]AMZ?ތK<{%ߖ9>_zMFpz)H{r7FmmL!! wsÿ$qY|NJ CpG7\},kNt;:G,QӶVN̐L(x1={|"b 6*t=bIVxw|_9\poY9H=_Ƈ5Pz@F;g<e7{dm 0 H/ /n?e <(+Ruj:R"t1ա2.TD۬ttYMhN] "%cbxS#(+I҃ 6Y&33ņt^|p}" r~U!]_Nwܵh%>$ÀsGE>͎Dav <%Xk9p,PpYL x/88-'vĴV ĨxR\RT|yvVA(S]ǃ3 0 ܂ 7k [P0EװZM}"PGܾI)3Q8pgӛ-eV ߹h)OVVDQ Qͣ&UzKn4 C=`zQ~7UJ1;S@SZտSSPޟ^z΂M-8q2=mb6 ò#.'I>ȗ! ߂cy8Yk% 1d1CJRF6WQ {7Uz7]r( CEyԻRUqZYgTX o5a}KVhTS&Ii,h,- Qmф?7UrLtbPi#9DfsI#cF sLj pM&VUJ|קqEP#ݠi(>OGhS)э[o=,4NBQc⟂evv\mzBh꣩>|C/`iߤ_=_?N j\w1cSNx/\DHDvYG.,gcϩr&̯)\<1 "Q!3fM۬$^%H*iQ0c: r{`8Dc(z*Br ʒ8/rO|L'xw-rȻմjfo7kS_cxi7y::3ߒvNdSDžvP2Eȇ1 x`*ֻx&|ߺq&əkMSo̎mt2e9{y;wHK^h g:4 pZ8Z|\d'8*zu5oъ<=7J]XęHwO#K gܟFA|jcreu뉄_#"q+A*٦ZnPT mAes!_fUuc)1bkq%[rP f?E$!lWoq;эΉcq\>ܡUm#^wE8C%.|-?w͖mb>Iu.Zy9Jt0}W߽HfϿ_лRCӶ^ZPڗx p@}MyW{Wn=v\BNSͧ5˵}9V17o-b(.3\4b- tp;au#ue:*J"g8jFDboT4JJ]l[n0r]>\GQV1R +RVh 崔u j4c&7PuΝwx?!.E`)og{W@p ɸî|g NsCPQV&XY:p뎑gBj! ]lri+뚡wL;0f2PO"XMGRxҔ8'lT7ƿz|"cj wGA:)2<P\| me ؎i ۘv&%řtP(,ByOʅ8 Y%z+;R4v[] A$qpw rl] :c!Ĭ"߿}m;M恀T\\h]q:E1:伂Y)' Vٽ0d q  /tsxnGLM[ƴk<͙?.'e *Ѹ9JQxG^B*z0>CZ+TX pyJN8r:Qob3Cv6[ͦ EM;VBX n,ZLk#. ̶=(Rx+Hih0)3Rt6h¶3zp?[ȮJU'Py@ @JAȪѩT>1Ȱlݳ*;ஜ :{o3\A3n%S/Io9%90;! 'F znG>?whOӠ PVd QE('Te8)3~6h}93j$I320--j".qLLO,qnb//72Ifty,W_ad I7h~u'Y3ecQ֡c8>eU$(:5/(ڛ@Bp{b "GEkg߶${!1e^>GD|>W З"cOcHdz2bcd@+}Ӛ'Nt 15'9($x{gWxB:{A8*j;}@ߘ܌ٰ |*f2v)Iy7y;!+qz;9* k bmh&/!׳[Xq.%ʠēT,2ʝ2a%\sy݈U>3 CEhE"М"T ͊  %=X))hD*%X_ʰv !TʙCba'3S>#衆 9knV^)_AgLEU(#'`N2coq}jtpI.4(*|Vs*Ey8)T:Wh帱!}Zyxzo:%c)UiS}:_ol-_f3T}B`7d<;{?KjsjɈ+tm\a@; cF|[lzd q-r={i7kڡ y iB%xf)7e4@yFm4A`ᔣ M㛇W s {׉Lh˾bvkOz)}:]m~\7K[ՖNqkPpd%AY_X4vĜ-q2_iR,'RmVXO/xn⯜9M(awE jdJL~&Ͽ5#)Ya5_B܋yyjTT}ЭHo>sOޣ7xna~f1yI2kXM\Pĥ҄t>You*2{f9 Ch\ g[sb=Xߜ/[@G6;8wMq Mƞ=)w8 ,)5#7+ΐE- NoqJwQ(>ey`fT1f$꿗> 8j~s4G:ψŎ(ނ{\,.'~'kƼsT%S&c;hn|oS5AYק /Bg41L6$LQE:BVJRgeDqx 1"Ida d@l$9$\<ZKD"H^_f56<_o7$jCzfJ17,l-bp]PjR?s9_ݹ }&+2y&KDA(Sja o2,LPz)GSrx`#U6f6Ri2vh}YMr<}dXH%zuAW@l#z]2q-iISx-*yC)3XVW{ ֠ PY]1fl# dRb.~ޯA&U[fl5xQƢGEǝ'ꕁ"]V[lV:VRAoH75po.z`;pldǩ:T=*:H3C(#(b^dWhYy4 Ck! cf8V[~a(ȭ`vIj^` S͔u(bFЙevcN/9(77ߎoGmٜrӇnMCхKºU`ۇq[uJꓴ d2k#=h^!JL>Rzċ]NxRN-Pxu'-wz ;h{*NQY@ RQ*Gn{}*%ecERuF*RҚ>PAydҚ@T~p?tAgSE&\qh˞$ 2/XUX1 n;:p33 `L@Oq/rvho0aj{q2ΠV_S*F L:{bAeUj )I^u~:CN̾:k 3µgfprfEA:f ՟Rwh z|Ne maQ@0rpqOXղӘhWn ݼme\Bf X:Ş@E 3A"C_$ OHԃݧtB&:29Lԑu\bS䛈O5F ]Nt 4ִ7i?y݌7ч?쭝?MY~nfwf9~qϾG{^d;~DLOMn`K[`/Wfv@LW_@fl |w߾T׍c䇫ʉe8=Mly/7ɵ y2^ i9#w)ޓi < A(LnA;s`Xp}(C =OH%I՚~?ھ!fC2n~‡?Hb ??He<׸$u_:0FAM뷏loǃ^"@,,*;L {P\)7Sٙ{忽m'~?^޶g0O*ZйJK$hlDd>}}㸽5 13EE;v{Uc߮ä':]'ѐ!DB>Õ+݀Gu:0MGPhǮց5\iNTvuM`Ԧrq0}J2n1(_Bw мF'ج~wE86|J[Z)yIn2_JTANmۇRL2wFO}Ѿbe&7/?^V픂}8u{ɶ)t񆇅*3SWn` NӠ_lg9x|g=mW60_=`Ce袹{i+*+k\T]㢫(n,t8yx<.RJ7/?S|ۻ%|o ?jw~~޻SqĿ))t57!-Vm poj1Jr3 . 2޸H][ە/ ?En:|Nnz1NM l&?;㯚.4r5HzJpyxz©s"Ґ{fa(b3(b*bVXZB 3pBIߠ"Y"cHdo׿K OߚSlҒ^7rۑ\;!@Z@Rf~9j+0g7x*sƋ'XO0 7)D+|-UTu T$lާfKގ ` la!vx׹6qDjiyMg\DrFee͐IsHr *C Mh猗 v̯S0T捎?EĪA92ED ߱(pE.Xz( ~=qe6gyĺ-t[pD_4ٿ@ks]W S̒VaJ¾MZuҒ0 GhTf+VƍhPu 8F`$d4M[A=!n."{T`gt= lsM"f1j=ܳv/rȽƤA-BBGhTfCDn k1U5,G2Lywe@v)̱ -t; ^6S8-nQ `o )pJfCQ͜@.) 3{Ӭ?}zʯ0me.$ &O\>8;>ty do_QhX͓4P0R?6@ΐ 駘(7u;P%¸6B[ 1NxቦKrx1l􉁹¥kbJ]>LSj )'q1G;ϭ~ }M.st9 (h֎UH67vHvuS"s飉qNڝ̊o(TMbXZ6/aa; vQ9t.#esqODpF }4*3G?-FQQ/ּ ˧Y7C*?B2s g?F7u,Š)hbx$v1 Y֪>aj9me.A?7MNhϛ>0_ě7fNSLQh iJiyATHt;/f,֤*sBOz?6}F]^if|t>IUV2fR+.ǂw84¶@BڃݧØIXcDQUΕ"?{O۶_!JsI[4h HFQf(Q,JDʎ#Ulq8Ǜ7wdp2 dǦL VHE2UwI|oUeDJ=ڸVұe1\}z{!*h8-$!|NT2p>J}DmQCwAq$>VQC5oL;*Y1F!WZ-Ǭ)I)?QޥpnA?$ 4,eH1ڹC  e޽8YKe{,`P+ᑹᨙv{4ݶ ).nk,T2pӂ֪qePETM.F1ZN?Y`-v0r[pQVV݉&Bw5$Պ:!xjIbvMX%w̰"kIc [tvm49lb[d-1++hJcr-鏧.wF@+3! ei\ouЩPjqM]gfv6O$0t\ 0YCsr{z$qޞݎ|%W翹?2;q`:>1v>w7۷~b#̧ףWh2}-h}:T5V=\c׃8~k<}= "BC"^/*O Ýt<0x')}`ÉbU|:/ƆhdԞlpߌuT`G̖ħ.p}rsq'vs7]Wu@L.#E8R'Fi-!o i#2$,I(fjU]۫_@֙.}nk%jd 2$ kLg޵þVޟ~h2eE/x Dt]Ld%]Fӧe#<&`iϝ!(/A@N݁?@TcL #ʱwZgB+Rf;1]nl*2xۈ1%Q(0"~ߍ՜.fC\RꌴQZ꾻aCv[ޡַ1TxM\K;D iz}%\j4uz/(`UpiM/_wôORv7l0Z{ٶ֙CWm ~ &;TO1BOB E5j)+5!tIv$;].NdK%tI/ɤ.Ķ*|Cq:YsL3Rj\ݙ=hA0UHE h6 hh@%0 Uu=%`hoޯ(E#\QeIM-r(kF`__V^lj`? Y .@`\C4gOh75@|;<;F+>`ݚ0ݚ2b4x^o$C GI҄") )"-@%4iHQj˜<)K\B⭯_-ln]'%$Vףn*#}h*czTa04G/^.*+uh4jBԐԓ:A5rYep4Z?i枚{`}g'yke74N6@;(-è4+ {Ox{iM@UD^!͇[jE],\µD_M| V,9&(:!)Kcy9Ŷ`mϢYJ5`}o8ԌB*vy]g$-ãZih2 R -?xhPđ©*0 NƉJ0쉹8Lk,]("'+¥đd(8FcT °HSa=OgER-txF,+&#`mWBXοX(C5Pu3'!V|6$U^ֳYԫ"Q{h{N.۲c~4|3GA$&hn "2RX gh g+éHp˵;Yf is:$:Z"%i¤"{)"* ڦVc 'FpNdl=brGPi;+6>u_:T'HraDI(N$4McŰIb ^?4LNXA,!FE\8pT4HNV,lc#2qlJ(BLTRpQ A2 iJRMGɔ(EL'XD(Q\uٵLjo0YXXю֚u{ɕ}\o2gm3g O0_ -jPE[ᇩEn#|@0w}wAZ"Hsp`~@nHg:RPn҉^ŃOl ,c:=,NvYFٵӥR1 .=R`q^l/ȅX, Ǒgǧm sx gFo?yǝ[$`[s[C=wszzω{%%]Lk/K;6\=ch[p̺#Õ.v߻n^W%P1a!CHS.jCFmw­Jm\~RT@TiFclb^H$O x*hx(K(8ImTRM燤aYI wjOaYߎ?ʜ [D[R6Z(J50t lUYk6Sgbs27bez| Vz/4KPWػgGysekEP O`뻈Y795I&1B($ 5E[kÌ34‚S :M9]Uwq1uf!18Q4pCꞁ]ß$ÄHu3lq.pxDSm$IeI(2R$V 8*u6߄u7&#v%`'gC S5~Z%ZNΣt'P{2X `R*c)2ظe ER%A>"Q@ĉ1g$Qr~>T֙#nBxp އ7exM̜>z`E6:vs^!qh?=!_̱xow(7b=c%;X藔EO:U T$?WP Ib,y$Aq izv3GkY{( Yp~l3w+8yJw.%;KkA2S.!kUdhDNJ2bZ ٟODȨq3OWC_zw  q@ C poz=qiϩsŠ^雃%#msc)ΊB&?~2ʘ,/ fk9& H38tNw qD*Hm#=>'t(vDs)?ؼN` RpX\ {]ldkV8\-6 TWgp l~x~:}/g3FqNȤo`ƩC@ X(s J3Ch9 39(,N'}|Ό7+ԅ ζ- zF1waP^h`Z&>`&ضh/{.b=A}cNhe[ibo!c.s-pX1j-f|պyTtCSmqrkCem(Rg-M#A*:%w _ojRG"ymڄNѨbsƢ.tTQ˥:ZȊB=YcDak9{:Ο`w FPg-T̳?dٲŁ+wٱV?5"u|q9ٛ\IH& V[zVf Uy9]>ujH5vIܮhr畝zMNs+:ԝ:{u od/.j v«zhr ǚtG Futk}aݗ/"OAp/UzԬDqm,t+':Y{4nA[yE3:jJl)/D((Uo mP EC9`Jy0xz7#s]%gx\gǏ3T*n±28uhX0zuB:Ƀt,z!XTNH Yk$_<{S9 lDЧcgVrr@RsgԧCBt r;aܕf|i2 ~Tq)u^fm .DqQg  $2pxu2UVh)-?0$R+HrcNm0-)pd< ,Ir̜g^,Z.2*TDuۨlN>iqbnW1„̈́Ir-#K%m~^w5ےDX%VUܒ$+I5C-I3oI -u>PY뙵 m?Ƭ^Kv6K}6K{wrfJoAGg3r=ZkZATz5K럷|ײ_{gEuhyJ^MpKoٞOawrqz*}{vz%7`NZ[ ?5O 秆Sp~j8?5t_+^ԇɾ~=Fkyzdp:V~-h;oGеT'уK-o6 @Z$.W 4O~l~'Z6T#>Б~20mLy =+%Ho=}֎ Mn4-vSVT5T,f&͡hL =tS.iS@9 QA9w ]hG"C0?iL=E97㚽4[9*+m_kAW{5MH[v0@)R{k"F5ٳ@f |Lѧt;7MVio Bf}QNϼN/.跕;ewXFTvd#\BSzn20A'X \ͺUTӮ?8\~/z=tj!Է>۔~j7 :Go>,ZI[Ym5'J/lԯwZ+ZWy.ҼzTn~=d.|P)!i:FbB ѣH2HVi8^61f+-wY0@ , r8`~̃޸.ڊ\"Rn sMI!)`m| |ha\ zf dyVN|HGPhֺc5 GB-.7 H«Ђy#A4f^(=B)Fk6+tIw3?Bn"2y jdh<9V YGPhS9I Qb҅N0 VBZt~ +.1 <0,lydYEy4j}7c(`^hќyuDciKmڛ8 -5SaژD52DX10Ғ3)-yJKҒ)-yJKҒ!Ja0 Y9h"fC &m<{  1`CR/'f~ V[3/ @0HN^3A_ra9SGPh/a2e^K%N@GPh $Ӻ)b"hYIYQ'7Ђys@#(bә4Cs1 4Ђy'ukB㘱S2Y4󱐍Z&3c(4`8+O+0x5ՍЂy=QH3qhuVL%,2R@%s[ځpʏЂybkjih?,aADkXC\!jP{ -6x(0?->GW泏~z Ef@1n&a$85->/.ǴΗ,Ϗ!eh,$*NkӘyY+ y Ĭ%'NDhبL:Ck#3ode^U \mMCShT@͐FlsF2嵌^C󀪵/5APxì$6-6PhĈC̏Ђyw^ Er'sE=1Z0oh 8 3͗ 3 =ӾU Gnt~ [a[CBQFb"dL1GPh|[a [( / Wh/D1 c(4`P7v s -m i1jb~fva 6gJ &<%/&D3p8JjפrZ36+XzWJ3{ Wݿp~W/4eN;}t{"Hӆdz^<2Og^¬HѤ0pih 3B4g>&2əɜy^23GV62t9BAql r2VBetg #.I=FF.Q1 1+BHt -ևBܰ9RQ4J7;B 杵u>r3 $`D  OM$q܍Ѐyru Q=w$C]N"!CJ4Oa&k\=F 2(|@Ph1atuH!_48P+1Z05`k qK 4Z&r,Dq2xc(`h:gZKsVq ,ɌC(CP8yjU xWFkp1+X.iG#Gf=cM/F6,u6 3p[ƷWݔYw׽y>{n:ECt[2HSVM޷D%X|r1G#0{ܴ!*>g!%O M`]A)$%GO^D2Nhk5pMm9fJ //Q42ZiAݷ>fAoj~E!_?&D1{e'|$1T*KbѵZt w :)8r9hHj> 2|:G^!>-.ɂ0_W")Jz4 @$<jEk0ٓ"dj=\}q)P$yCY&ߒV1os0h).r:kNp%>͕Q,m11~GZ)oښ4FEj1\ %FB_T4." :G6s0I*Ҿwo$ z3" kr/Oi9p hU&qU|lHJ$T#$#,n9}l~D?V!|uQo/][kdi〵7JOvrV)ZV.%F_2Rs_57aI[_t O'4ɗݛg}8}<qxol(zNN%;/ߞd>ws271,*E8\z%n?zoɍG~%mqcϮk5nrg*v',c*cյ4JkmI _. ~?wfgѴu%$&EQ)Qe;vO,5UUuu}Wv0S ,.Ka2R0QmyeµprXUU~χ3;]Tm4-{(^u+H 6zun։HWw].jFPW[sej~x4ah_yyٹ Rշ]*2y_70p#r Z<[1J}}Wˁ#ּڭiUi`(~0ЩE*?'i$\Bzi2fwfzڄ1b2Wg-'rEl$z\Yb9-SHcDƯ_L+{4L0ޘr_e0J :+6~UoU1QShԩtY;hbFxgDRz.ɗKUt}=+|̪&V:&T&&1$`&Ф-E9Ŵ"B $6,ڨȽD@+r|lit S 1Iv*Jg_XSCV+r'[)moLH"E Ơ Fmy|-LKXi @À$S~r& 8I$Ynrv L'VB!u؟ߚn'> [qJjQ˽f,j[J}3V^z:m@_6ݗ E!W[vur~=:2↱,ԑR%QQQ ƙFRL<1'~\EDLq HXlumR!DL#$i"GU 8Uip3+NaMi0XA1"XØ^g7Y _-l[4OLΎ[ տA2 ٧&ޝQ(ծʹ۸8zR ^Wly&3%ГU2䐳|d?~>ZWQ-J'Lߗ$\"84áu]4Q(AwƆ0A\KuLf_S;=kWcO.JeB82$֒i{*ؽ?[7%-vqj]$NFALzކE7lxb[VF3݈?s*ś*B6g'K)Vky]S tIڳk[d_/KVT>ę>NOg3ڭ[;ޞ7O S4!ɰ'䵙_SM<ٲ=㆔p[\'e ->r8\ nOgY ] 7`bsQcaܨ Pe (EDUzş [@f{+`?G XZQ֘lMO 5oޮlx9ZfZԟoӖ߾L>e6]^n\o7 0@'v aotv1U [08N5C\`ۧ+e(zA쨣o ~7ݡӎW*fc.>7hףyGEMz-((cWfaq@)V07OKC62IIB($R2ilhS,ȒT1m14K}M{抋; fvNw&su_sYŁaBV<Cb\S} ]>ѧ DS} OA|h Zu7YZ:@`tMQ2@x"(eE3k'cS.~f}PnXW'z#q}ZS^x4; '}lG9%1R"5$#3$U,Q8JqENL4E=17Gim4Ӛ6r;h?OXckvaXrId$0/DI($4bŰIa y}%~izJڄ:gLFS2 ˊIp 86J%!&E&!ukH$0B3iˆMr0Ɍ(E GӉ)Yz2`yk:`%e̍^L/ֿZP>fXbPjWv  Pzԅb,)1Wj~2]P^N0䆗Oxc%~ٗ(5_/ʻW ߯vۜU9VDU;OQ҅fn?Le)AK9yq$_.mzFWԮ1XR$VĐ{hx0sC K9iEHmXlQ{6eW<2,}xz񱼨Q;|/-k%]@G ih [JF4CT/QUA2"Uqsn#u..Lr1}xF6^s϶X2nehjM=!&Ӫ5Zz%7 V`[$sƒ S,eyqI848IZjSeI>hEP/tOiT$AU8aaMzXL =[N\S1d "DA" 0Q,b5 B̒Lcə3p'Ytd13ɑY(3X a~kΞ}~:4բ8u*YlkR[3V^Ik)5{Iz7xArՏ^`Nۆgy}3BcSÂx^e0>AO*pŲOM8"\!W[ur~=)7e C$֑HM4RLiˎ#nqyT Z[(AC*?1j̯n-_2 a}Pmwg;sqGlI >p?L]Z|[,rX~(6Ub"Vfa4Ei4O-7O&7q{d]p.zE_.zEO)|K`X vŔ:T exv0=|kӕX'ܻSϞ8Uh&}L&t ޙy/ox(,Õn?e(NPLI.Pţ.?aބ_?R{Gd )_i;Ҫ+JҪ+>GCcBg1HϳzS,ԳzHdKHEJOꡔO4&ӤzTOiR=MI}47XsOy;8?%OSbGG tD?0LSiipY=,޳{zBY= zPi=ͽ4{{%i=Ӈ4^=ͽ4Oz_O='ĿzߗL =(>b?JM!k=ӲgZL˞i3-{eϴ왖=Li# |08E4-cb^$#%1D@ky}g~茒hr~%SEJwG.O­A>oH fmMGu].vtX= ^`Pj+ObNI_L7m[VK )=lSOF L, aen~𶲪%MZܖQ%@&y$˨)!&!&}"?.>ߜ~OM64 {B]o# SL #WҷxgA6"*v=>-Xyٛ~sV"۪gZF_AKd!*U[UkbԒ_!)i\e"@`fgL鎭Ui1qxV{~Y`(+]WZHL. NV^*5,uahF@@ixu_տ5Ŀ9ib/&zz8ftt t$'t<;.<>$=\ُ#QEdztHMzb&-4CA*-YƂP~Z~Bš]*$TɂĔ#"AsTo:~KaGJՈ" 1B#$'n27a"{&CDD /tWphRDl,,xGI4 Gzo9`M )n}p.#Xj;n *1](_H4sGIR~>xd1EAyQAqpiXeӘRӞ7_m3#nzSÞW[6TP삧f0T`7f;;:] dGYf*֓'KOa$*8p)BSϘ8P=t"Q}~W5(޷ۤ' [4AǖfcnmuQt0f:L3`\t95f:0f:L3`t0f:L3`t0f:L3`t!m pҶvֶOB.@S_*Bm(B O˘ rkN7MW[dh-/"(rֹ] qtjS)Nrw9'vk"WP0T+6^Ԗix9FPcib A_o &}?Бmӳ i,ERrLU@F(e 걇X:;HBB;:p>mOEXE 3J`>֎l\;O>Gz|^֗Իo/-%AաˊW_+1U=nn4O+g7<%")qJNVjMz[˳VwIoHݎ.øIYiI*VB1ICBO'N)╿";<{ƯOz^wכ;g87x~ͱhlC3k3 L58k|T8(U,d9McM4V~ak?~pu'ӠˈO'@Xith:R?%Kݟ.QTl,E `872!4ukB6n61# HD8_g,hIz k{cD,ƺMډBLK6!cJR%c6(S m1ɅrEV`PT\!ETx*R5,;z!tDuŔ[ko5{ܜg+Gu)]!J2{.j\jf}qjG^ k^WRacj592itDRjNS#vwh5L'nWa]3 L܇.{t~gocMCȠ:<q4֓/cBy0Z܃ŵ3(ׯQǭ[9C ͮH+`!& .0kԽ0e}4ujӕ sUXƷ^q3DfGwT>36oߦ5UaZՒC2_ 1$AÜLIAស˛ΦiWim%f侹RTᏋlI8o \-j&n4K9x{軉w3-,g\eSvZ+$w6 `ϽVɆG?z-a)KrWv"N[/FT ,/on DJLgW4 s0(D) Ãli%% lD=/+S)Pʍ"6oȰ, Ƹ#aREbD h$KK_V'vNc7?Zl]A>=iG:\o>##,=Q| 0;ꏤo*7WjM$BV[7͹4Ys(RFR6śTb=4} =Gs_~*FVst/4%&%FS" MH|Hvz_T=[iewۃ#"g4`[*$TɂĔ#"Asx(@^ e`V72tqw?džc ntGIR!}@c*AF^!OPNcJM{}$rM{^o@uP|x\h.mNvvt:eG9,*֓WKO $*8p)BSϘ8P=t"Q3~"W5)TmIFG]]-G66:1fw'k% #)>Kŝ߾=)TwTZj:Sf3% ֥07KO[#9-}%=kk VZkA 2g@Q3 sd92g@ HkHpIN+'uΜԙ:sRgNI1NT{K9wZ:{(Cd1q0`ZFC6~w-Ϻ}j_-;ezuP%GWyjq+EJu0Ie%#]4^`rXxb GizB:b&LI8F.UC4bµ`8z 9lZIb Q,ҕ`ZaȖ ˥m]WZ.m~tWÏg-bY̞;}dIJ\:E4c!VX.Kr^Jd0' RZɆ9^>AىHw"%10( JN%10#0$(qBк0hj|dŇ]_&VrIdI.@Qb56 D5q4FJ\ dDe6MeL$f~m 0U#uR{b̲"8l [r!&Ey";vfE ۿuۿwYVS|"{zSWFK[MK鶧0{W}o|Uo=M)36Wvd ï+;UHcDֳ_®UZ%ejfKX64>V)Hdkڗ.k_qup'.{S&~.w%tiV|0? 5Y$5 W4I,@jC?@HMIMjH&v&%}x~~v{}PԊ-gªdƛ>{:̼jd׼}}o/wi_r*B3Fk%=/%vEQ+M jT YI|UWUW5C6o7Sr2ȳym=5bAʬuLBLJRְ)ݵy_Paִ3霯ʾ߶^>h9_q9w8_ŕ c,0͛79ܝ;P46*2KEo!G!=RrJ5Ɨ73L 1bt)QM) Nb縷4ա,!eΚ:.͏u9 S1mBfjLhb)SZE;ͳ}9D D"_c*btQ#> s^fGm[{M'Pǝ'p_TхAUI$dsKQ4{ă+B5( Zۖ:`v6*KoS<.4P X<"Ԗ"x $xqnXkm#I_e :ّ $wY{ol 2cIʯ}T)ERC![tϯq.Yw*iʒ=# Ly[9$s ao&+kUCʔFj*e(zP1)&8 Ym:땾yuvܷ yl >WJYI Ƞ0(ʩZ;I1VDcn !UVRGR@$ x iMj ncvFXH0, uFl;\aGN~'8. åh\IYbp Ao"ksO# z2Fu$P&On;鸴r"#${Z#sQL nedl\6r~ֲּ[ԝ X j݁rJ)o 9W8.fTG^* +(3xn2yr@0ŃUMPiEŕR$Xv*hmB "Q#7)-]c$ˬd(FR.|3L#u9\̦v<#_q[1o<*=eOwOˋ3/of=("/f5.ͨbS F̼.5w!rFA*gRbr𿠹oͻ~X8KaGJՈ" 1B#$'nȃ2d =!"}j4X{)apx*fR[s$#5kz81\myz,:{ž6]K7u=$󺬲D.(I*1 hLzdAP\:"UíiX/Wx39Mq6tQ i}c-h ޗ BV1$beSM+PXt75:<y,![nu '1bePD.37Nl$*8PR1qV{tEڣZ'o϶ qkuD=i;@%Ŷc}-'q4r6m^`oUzTbT߽enVkRE yOP*._);'x^p hr`*sVA-m=3{9A Vpt4:tlݴ>oq_/-|c(U_⨼Vmє%NםzR6zmT!蚐*C^0nm8dPkJ6qbSH^; kWP 0ρz.~ˆ`PKk]`faZ3:1S4[BGn›"~64=o_ al(gRE+d%BSS5{ >Dٗ)>'=o?eCLp>,M-EG)^'^'^'^v7soD c de$u/#GCĈ gKoD̊Y|2fzƲ@Y-`#RǑ҂~Q1fq3pt΃/rF16 3fܑ q|9~R?_]ݫr#?,=TdWtTWGgnz:J=5ߡTdIS*AՀ9dW0ś|G./=P'=@n{4i5=^.cKw1˵:\uRq2Uĭiƚ7o"'d: hwfuOw/~(mQz(>iP]5t!XJTC,ZGE"G9!+p4JazP}=i,g.w<EN[S}`:oQyƑ.]N9,G2J̃2Co~#=(zel]h `&0EN͓canhRK d 5AvJz(0|ha݄epJ.!E~!J,F&hBZ)YLa󬷦M 3G|m *#uR{bdY^ J9"V4 ) _KVgh'1n~ΊH7^{5?-A>_; Q)`o~u0z1S^؈?1$',Ccc,W?oy8p\} vƈ"g7z5Qo!-\mS0w-/r~MIq[%E?Mߥ(s$@o 2|IɈ}HPo./ .O3| #%O樦xa5_Q7 8#lL6f"Hmx<ݍK܀_tVI9 8I,@6$߂`$'tdh,gk|)׊-0`:qYm qڍy_&+CpK+/9c#JJJS]r|rܐ)D-HM5:z.UhhΗI]XBng f9X$'kĶ筤o$JvU7 ַ| nG]om{'|OhCӴz]QmMw9{g/=mVBhZ[Lk1M7j4%,L9 XobN8# ÇO2?A?7yfeg%@(zp -pe4'6Yr%-q2\SpI30BL2Q ކH8ƉCH.9s:nx1X9L0Gn>|/̷NÅAK SVy{~̆;ww׸x"Zrf6PȚ7w/奢pU0O\TOCz y)Kɒ-{p>a PN60lIWgKy$"kY"MɁ'ŜVk΋U9`\hBb,g\jsEҒYdWJ3+ #I1VDcnw+ҷr<"a ăPHk4s4 g9H5.P D"+vS;J2Xőp.uXER%V5(.!| A&S1,אY>Y؋1#2xr2˹{JAk8h5ꀩ%08VO?kYCOLжΠ6U@Ll܁L^ȹE roFo ÆsmCU2Ź$u^asӯ7aח VA @0V΄bKMZ/@@C\-]ҋnvtb&1*#/溸% 0j65i 6sdf.M#<[n@*ǘ4x4/kI Z<Mڨ[6CqP&A|vۖV]{mqۧQݞ =):mԝiEԞhVH}4a7WBm%*p$@[4A[^ڪ5oɂc^x  XzЫvڶ8ir*bz}zTs =(0of@j5 ଆ\PV He%f۸JMm<"̼v8*f%jZk=U+7_ J)F yU3nJ1iKAO5nJnCmJЅnnwI{J\MDkmH$ۑ `rN,[)L2IIo̐>F!))c5~TUW_!qf.P ma T]`i2a~du;c%=rwQO% LZ#ӗQ.>dU-M/{[")xH I5Br2t_%>?9zDРBi&on&GclkN\15+/^6r7P+R3EsZDcckkmDZe**ʂlPvߟ^gÂDR^FEN4 7t,qߞ@v~tO4h{͌oAA}x7amicU\ZR"7wʆl0Js0sif7a"˖߆N.FAxwIl~~uRwfdWRLeGרp:]Sٓfk} ƭQPF[m \m]"[l;%XEDI&zA?ۂ7eٺ.xSk=B V_4a7ek~+GL2$Ąb!(.a0|.ʬE(Q >qfƌInP}h< u $7T&JQX~V8d@ͫJ|ar-yr#QmGsX\udXDcZ)"V_"D4WC<|paKC-p8ўyA@wAQ< znm `hn)1,澃|]Vf@0[ƣDPM %jDaB 7J@.(jV ,Yydc"k/%.D𴗅,s$#-gh:۳Z8N^B6Yn䩍cjwL1Xa>8W0S dt)&쐫}@#x+:Pσ:KIj>-789(5m9%͟x+B2T)\_oV#[u%3v͹U?bS-.`l1tnUVf|_477E_i3(N nInݫX6?( TQ61CoG_ۡ~؊!+30ga`Kvoo+ :cWE{q}?)+_֫׳6xt4Lr]&_~M;|PzY5d~, ü>eUzzWnͲceYa?]G7٫ƭɍI;6p&Huz0Keg@d {ԳEhi 54XUy%k(9 2ͯm^&_\߿zAzeپ>E))NqaC̨R-x!!1@/ ht:3&<"hfSER5 u:U U<y,![nu '1bePjՋ[WgNÖ IUp{å O=cF@+ӒZI}eN,ߺƽ0Ds;xLC(K" LO%|궺 g%AA+?5ㄵm1 r p3p3:0cK rH# be}0z)#s0+ ]c"mܕ~@7S_ߌwUM^>mMtj'%GRqmx4VYE 62.wF-uQ9b0XJB=2m1+zcQ5p﫚28^ƴ<~9egqʴJ;@t+rS52jR0%¦]@VeY1ŋ8}x#_ger/x˺X} 0"r.|cSe4)qHp X*ZC&h~KyW]пzeY,2Ka 1W jʹ(^Z ?]b&9>YC3`9O7R"gmI 9Zr;_n)e?f|CZhHGZ:m tZѳl혏<3s!BH9'=cs 2X#%K5TV2h|?H%%~y$Q@u3 6, rkɇ$8w)k~떎?mgؾyvؖO/C^!yZ [="s3[/nhӑe I>~vgIx eP rR{T]mmS"O&d@&"'O':I9־%FԢeB5~qY):j,9Y &*seYih^=^}Ojgj5,j)WIcy=j؋A:m>0BF( ARYW\+鿳b5#r|M5M]5!I0?n$&UYGk!%%*@Pv.Z4p<1pAE"e<ջF7_jʺBUf(-7J.%A|Q+M!j\ Z٤^,r@>el0X>>Fh 뜽]䬃5bg=UbIʬM3CIJhOR0k=I|(%ض5oL7,7-Mt7q\ݸCbyPQF*)~ plpQ[*0wHs]~L1C'?06ѥ{F-Rjqo thB{2g ӝ|pbקE~s "DC@FQ_c*btQ#> w|M@mԚN,=W Vۼ8?$Q}oS5UG^ͤHrAGnVNlњas'%oi$O7C5gxkW/\hs~5mp56J!۠LnV )&.yZPT\\ H8Q%7Fk](5yZr*ţy>suv`]M2r*.<rÃ8qpYn4S{9~a3a:)9GGh :663x sZD)3ln%9JypȖKd$z|3J| ghs`cB)7(rnӨ#"( -J))$t32ά*3il΃'lI0KЀ('h'-`>53aS~NV(Dg'#,e_2%xI8# u"JR I!rq(C"8U('FLDyhAQZ!D֎hK*dȈKs/X vXvc p| x*œS" Y(FBЂҎC IIJrtR\)].WJ+˕JrtR\)݂Yvo#5j#%\ai@P# +m[T[iHHHBXGG]Cţԑe 8X8-FVSŻ0&8 JBa▅M{ zKWv t].]aKWح$-n+kKҮ+TV%Nb%*Abast47}ij~7ٶȡ*gRbr𿠹⼄X+/q2M$:jbo(aT# 3J T"+ ʴJx%T8 y Қ)RB%@,[Yh6LxZu8=E] 6p#OmPcnX8y+?:oB>~>;8Jx hL䑁zdAP\",BLAil5^V^IM՚f?%& Nuy .mݚٙƬm߿ l\qu4hfg.U `q:U (kYLuny$XLjAu*nTij IUpҽR1qV{t[#Swn7j Hnfxl9P=P֗EB LOe:7{Ǎ1"i 8m q{?b3Z>~{3cL#p,wz|U$$4jtbqQvwnlTD0$] M*`⒈%CWJd #[U/Ǔ4S10di MFƤ*LBX5&bх(U_rұ'c0e?U_=@f7ނ֝'t#fb $r##UtZQ!]ۏjNN/juAZֳ߀8⼦"6.Cw!T"qxjl: i="Y=sa}cau}k}kz`_׃M-%¦*'\tҗ#Is 1zٚF*rJ] 5@&ۨx~^ 6?^^'}Á\ٍA{[J)Z_ %v( ̆!vKO|ЧgqM.LN>ٝvF}οjymtӻ;O ~f㣸xÓjW|@8xOqlxݹ9"󫊇 ۏ|vZ7J1 YXXjG]~(Zq<hsI<aBp?Rάs\ZLN O L@oن("X!۳Uve8"V?ʍ7 ,FwxLIZRТ<ճG'G\+9JVNg#Od*!Vi#hpi 9tcZ|Ox=O\ϿMٚsQk\Z*VIĻGCɮV dBk/9Luv]˰/: oe簋id NWj:CVBǨjvAML)[峭@\S"Y7 Nv`/HêaRtF`*$@٤2"\bZ(!."YeұI^y'`KŐ:[R)R19*\XZ]\JVRjбiMv5bĭK"̖wwVeeE]vX=;Ifeqҝ:;'9vFejaj7-㿈ТֵJkdQZ+IT$%;RU1ZQU/?csqL]]/zh,O|VQp&Fu11A<Tʯ Q)w\,IH ˿߳vsN0ZcC9ibvUd..Bu.;qfVNZ.O jQ=ͫqo*.? ۶ü܉\߷{jhm>D-?^\gBLtޞN)[d֘D8ˣ<9^zW75ƪd_э'/g{oe{W7~իӛ oN= %}LhD.F=IN?Yl݉i#( |dr*Z)%$:.xWBZZmʊsI_NL3c2 s\Z{}Orn25MGҝw9uFEL*Wj4-ɄvVr/RA&ClQcnL%25ř0b6c8r.V ɥ,l|w*$B*'q'Jr<鏵ɽCz]=(nѬV. e/rDFZ*Pov*+!bNcgxN}n24A^mo%&:#Sh[$쒃L IBW)k`I8wPoYELDq!52UUTGZ[C]rNE`˨rU[nLjw{ݐݍʿ?vtJJKZ>GO9#i&~>IkHG~]AQ* (Ћ2z%tzֻkQ%E%0B."*ۦGg.X0y͖Vk ۆכُKob7;omO>lU\|/`-%r9ó7K_zi}_z6?]ϖ:c@A~jp6߼죾oN/O6?{;ufu'#Bk֫/qTCXs]G%.j:BYT?jmnlK]djA+^h(^fV&mEF&qQIKUsʙl"ZZܒъMK Q*9iUɐ F#gp[ȸ(vr٭xXm[Fy')>{Co:W>qODRԺjƕK]$_t6V_X˒\s Q%/ Z4)(ʻ_.gӗz8Ȟb1)ʹ7t ^ mSF݌oiZ]YgnW_UA" a`<n[I |٦>\[Nke+g?crpܝ_7}e࿍6>[?]g)jovZyhŮIEIr+;6> R۫:6[f.!/mcْ޼2נX>!2m[S1Ө`ˀ>MAԼ]/]j5/3ecrWAællїUFjtD=ȨY&6]d[ܾLѧ(# 0KAjK@0?^/~Z?ۊ\ E&!=WdENNI &i$ΰ\zzC$G+BSZ|2BYXK,Z9۾ ݢC6fKjYvf?^|Zj #1jnu}/t RƤf[RXʪHKIQNB4=r,#!23VXB1I5Ǿ'%J8_S :yܮ݁=-!I}9QRSBJ;ljm&!M*C: JX`a4)!s1|hh!lM@1g6I+Jm@JzRCl@OtK>ߕ) g]HR%1ӚXZk!V`H>q:{bm" ?n V%iEXB ;' \I 2kFz29k@>Cލi[sRBv ESpśN*~RCzFYo$VpLjp)DP+5YKD)2J1[EAbO/&v$ N&𐔨+XP>I AbF-C/JBl%VrV Y\wPAjswQVm SH*,e] "bOcK6·v ( :Y-`e=u+.U`;eq8 mG$Jb(j8rn eDUPj ƨXa,Ҹ2;FImb\EV v :ZqUTIXI(#,„6Zf'I1cVbvNuHk#;e&f@57]?  c)5&1jfl%: ;£*Kd ITȦ䛃C$:3( $`(׃Ux&d-X)v&%`g4BQ\@ᔝT0 A,:'l%VkH5=*(I0̡mLCbqыV B)AAh 8X/{@Htv+WX  Q+c+: @qck1H[Rb\,|˘UpNE&*/$%!&"(FؚХ{Zm*\9'h=4]SwT$Mump /`0LP8ҰJe%ӞKsuAàܢJ 78c;LhTbBG3!<(9@i8d§eU ZjrQQ)0dƵ-SKm`_Tɠlc<oĭ lh .?{WƑe 1OSa!X`A}JeJ!XyEl~մĶDvߪ{se,XdP?># <,bN$F@&/KA Hb'"҇ex޿ܞtz?Ywu2$6>P]( 3WXkp`a(cP.v1P/pC*U]H_z#X< ^HPAʀvw`VڞGHa =Z31 hwXQ@*f@c>'xFRʃ7DѦ)!%6R fxX@m#Q̀#- A-1Ev zȕB~)Q=2`KrOYԓ(w)XZrL!2s A찱Z#U՘u|vD,0IJ >:zf9Фd>|#p7@xa$Ti@e*,!t(OW8"ND42(BU"O|.[l!!N$;묱 mSkM}}WR|,;zyЋNȂ֐i̿fs9u/֗f}ޒO?5-sm4r4--'qq&q$Ag_m?n,wgTMw:a۾Lʟoow?ɳ; Zoקc|`3%`5籹7x&X_c?ALQ»4 x{(egEo݄yD|7j3r'8t !Ma[~B8]wBsA]9mYAx:\7X .'KJhiM \Nx"B8K6sRaK nmM{FW}grrj{=,3gRm</<ϠR-aa\^J[b*4%,VZR6E&β5Ew{_r['S7ړ;a\*-hYx} UXs*Pޫ-o&w7j1We~Lҵ\sz9h/ tr /"מ- iI9ꬋ+G] OHF" \NMMW/p RLЏ H߯_6k/D?JeK?MoMK-.*9jzf;&yI^Lȷ>y~ p9RX{>f/yoxoq>2GEC~,idwwÜql|!vDmED;.{ў| ,gO?,uelEǪ3ڒ%ڗ}D%K_G LZ$/#ڞ#c :ICW$+oE{eC}t&bг+_?2?}2JټE{G3W_QoVhohsG[U/L]u~myKgdiF܅ϜF'Wdv^i}Y$x6>v41@ ?ױwsq] e)q;%?— 2H["}vر(opv`񧖖kg3Jd!gh=tm瓖S|i$FZ˳>Cˣr[rv׫'0VGzA=Pe t}姞|r jǥ7E~>;笞Ó1MNR}NN[tM u6&ل:PglBM u6&ل:PglBM u6&ل:PglBM u6&ل:PglBM u6&ل:PglBM u6&ل:PV"5b!"RRK)&R) h_Mj '`Ux ZBkкbvfGxgmSvRh9:Z+ӯɘz~ԗo=?_|F6?d!b+GE< j4LaOؘmrwiVFɎwcՊvTVUqܥXɤvA8֌iM>D1ɎWrĸo&'ݎMd3>vF-9 b )|Nx:< 2e5͎#6K.6$6iњ[WR仙wtz"h/}Z>Pm0 վ -sVt@;|ld<;aq#>7Bɩ!5\xy=4~w(J栊w@&:@qeϋc`kɵ3LN7_uwqi>i 'b Rh{7\l;%6 ty{Poټ7U`Slؕwxlv'?#/bϲh?X˿Xw7ӘZ V9BDj1P c#6@, Bb1zG"p:E;!\v, *-HT/̘'-jd2r•!R@rIRADeY̆V4!;@)O jjWxX u=O6 v](_cj@o]&Q?@2N@Ơ) qJbXUfH0,d@^RE۬@=-!3=Hxb"ԕ24ac?%A'D:[ L #, +In˨Lf_*d%E- N/ DC2GQF4&A4E<g/:ZTxR͸ٕ(Z3dif0 $ )4cd%定%^8A(8Ek,@LyhD횠 3L5^VcbII< L{ eo,z%Ė,,+!Ea&@;,'JAnwa)a'?"֯q#|8ٻ6+W}ٙʹ]u6@ #@&PO;4!%{A&)RRSز$48S1aZڼ ` lׂ%a1'W0J:í.8q-ĒUZPt(;ɗTXzT9h7E$|@"/ 2 H S"VhѰ~$$/x43jP7fDl^%,:=(J@$d;I`~jM@w"4_XfE!W:d NDj <󰾲 \{# CsxtIKB4A F0X ܗdLp,{eF^HP&")!,ԃ1:h6X2ZypyWB# 2v{.X9(f@b|N^ :{!0)@q"eQ82z'emmTgP'0$PU<@-Ho2"$Rdt<i& }P( B66r$К9i/t!,-+,3LF(8p ̂+fT(2IYWX@K+Mw a$R' G|#NKRЎS_XyJ[A7鷮߫i˴l>W@9< &d) oƁ!Y `0 خ;{;DŬe°JKYɹYi`4vtƨ@ȟφptPac6RSp(^"&2 Y͑rtBST\"|PD .I;T@=L   Gpz V6dAv+ J2,7VA+DO' YW"&ꬍ ɉ%՛ ط.J |P0VҢBH.|yTIFlY+zZ),@a<hC< Q`-I{\ĀVX' D5h*=|0~ׅ0p18X DBL%wP]%JuCCމhGVEa`j?L/[J#_yvًg=$r}݈Dxw5pTl~fTQ)>E| @l<ҋFȂi >UC~ۛ#dk/rϯ|35``zs'#h{Dz+iL^o6)E*Y^۶o y,>t-mZ߯"u:^W .vaRäݠ~iWmI)/ρ[sV2E8$+2)g/mcq|ò/xZk7޻3-/niWr$}{O*^'_&n6E%/U#?ɽM"yʓTHގ^pbV}8fbRQW'n9&&yχNMz<ҝە~<7KKé?}겖ujuKnm9k/I *?|fψt:Hu䤹"$W2XU⮴+:{ygG#dibsrHv//?@;4hGZԉ rG{$AYSj6CG,egJSFG ZӤQQ eIG֎C 8OtWJt\+q%:DǕWJt\+q%:DǕWJt\+q%:DǕWJt\+q%:DǕWJt\+q%:DǕWJt\+q%:v[nەV ͍eC7!G>aKa>':cG!.φ Mit"1f [+tP^c@SJ"B4"}cP.:?8?/Wsu{H@t1笟sq3ErZZӐUIs@0+@}ܔHd]f^o3AHK1b,Rl26  bB9"X~;~3ۑH2e 4&aG![};EWO0^zŮV$cI$d/)K6)J8"S<K8sj+Ch+'AuR-X L1XQh<%WnKHN:a,eB-dvӒv*F4 ƞqBhAj/!#!Z;ejɡ4m$D(h'dc SE~?stI^%Sg$m.AqNT 3fqhKUNJSi,noJĩoFKO9YQf('yVfˊG=i](m .oT>p056Wv SQ&Rդ,I8 GlY׎ǩ(/[QO {mKwvK=9㛨¨!zghn7ID+>@URyYjpjZϟТ _N={3-[wtwTE*`Z)J jMRHаBYWKZ{5%-8=vyz|AXQBDhR3m Z~S 4!X=ߏn_nocX^G\|̤l_xqˌ1FMi[-XvAO 5yz?˪b,'Q`˱9Z]txHWWߺ»ܓZ떞XĈ,~Fo~z]qpˡ!Bvyotsjn-kr֍CߎVŲrҲR3SvDž`Z 6l]›[ӯ PЯ}!<,$H(L3KK[ӥĽ{ߒ{/f[dY4Im_$A&!>ެIlֻIm9Ѕߚ-%Z^Ղ^nXÔUDa~I2I)G]嗼?vy6C{e]RUZgc39KX 3͐]Y~斮@xQ=_.|i'8fgޅ_2iֹ$Wl|ޯw~{?߳KY݀.5˅RY,[k׳.(޵maJ/#׉c6oKLvVv6z@UX*,%e;{VJ? NeRd $#X1\"_`CnqWBΛڧ\ls[pOi2zMe!'Vh6CENؚ7tQRJLSZbKb_X b ٔHƴqUr&{c-]BL`1+>: uE:8;P}{]מӉ'zsg|}}7[R:vN\>z6-m&Qrդ˫/46H<urΘFBk Q(亅J!L98JH68kF'2LIu@ڤa^X 0Vy-à8iIs;Nꐞ _+sހQyNl.,| ޺0.kj׶+:z$t4켨Ϣq LDrZ:j1;۫@|T!hحwdD]@(ORFޡhkyr}z4]$ $ܧQkFOttJ8wSb44t5g1~S窢ݭßھnh! 0_#aC~z N!ِq5y.dEZ=ko"I]W;K-iI3eZ|ڌiݯYU`0m$̌xGs/v.CHb篰عT by.vb繼l.vgJs!\<;s\<;_B!)=Z1lqU`U+zMR+)O`FXdVy*iƮt>D-8Gdَv=k>5M-T -kI|mG] ~lu"I K>ݏE_q-ޅc;0E+A-(bxD!SGj)u|eR(I#4CkG*|D[B2^3̰"hy$t{!L eZ30{-#`2 i8-w|FYbIwoey;xXXKa9n2YWpKy[i.,LcfwL]sSOn9a~ȭ&&$Ahi5{`? ]?.ۧvL*nrBYȢnG3? =y%b楑a:lotamp~2GwLMT4ë# zK`χu=lv^09[UzIvwo/knY7?n.Z8nn|2Q Yah `H[YK G ty$$w]]5e\a,Py`ap{POq S҃db\*gY{c q46\tOfhŅKtA$K?#I*ى! b j-qCvY@iI(*uW[)r]q#5vo,We:њP2zEx#h+8>jK ^Ȍ ^`A ƜT 搩%u1b3HÖ&ٷmj@aFɗpvBI|4 j-$ TH'K -"#a:0- 9|d.(,r!cb*0S@u4H &$ڐzB!c}9l4 {db4iјV!!3uS/N Ij$fXǺeLiҎIi#EM0y̅='+EʝuFٛ \R1A[5<%,'W`m3djh:ҸQ1\]%fUW[ILE8v0K×Sϛ|ˆM}Uxd!R&R/5eDDL ` Xy$RD;Y;^;3چz@դp>n{Կy; Kaо(x~(H0KwYL7gI7XH>?fhp %JO^;d%0l\'f3|梹K:3uA4J҂ /5C_eQ .LC)sef̄y:\Ӭ36ʼn +3F&J\^eL (ݖ{uv+^,kyhw8/Ȏmw\z)A]5rt˭M #W JR|q;:{Xx}dj_Q6wutU9Nk3GN՝5lPj- ߎxSȬBFxb#eE{NN@t!dWssM%VeʋfpӌJ[sE6Vhz1Dn-,oa6[(D .8!–- `7-L(a-,Ho%a-A>a-AyNX{ kT;;mUINk׍F5-e 6h?+dC$x!k%ʣVq*Z0E -} 9u8`S l)ÝgˢD\먱(Bdx@Yih<^[K?#DЦEKuMƾ&N=ޏ#>Ր'ߠ!!#%GaJGhv:0(?HFy0n;`r)  }_fyEmYLJ?TnDR6@267??~)q`lxV[N!EX? =/[?8W+c]2Xa? ub|'2?k}^PG8* . _`W@#:])\^C3\V0XЌ}=B(_6;C8=kåǼ⣹?9ULT#LH[LbH=0 fL";Ie jPWx8$GBݧۋ I sm݄oh̨NX;K>T1)U6Cǻf|@#Jˍ'|J쒋IЮ4㨩ƋN"Sz$#-Ֆ;ЦcjK`4Zn9&k9ϷplbY8CӀmLBL4VOY_1)O}S k6XV[$_m]Ǘ7I~e.5e5nJehB4&10 oJtyu- ƘcvX/Rb]uu+KMbN `fx{ ꐻyhf}bu 6Lo;kH)EВ)E@Є~J(&~כ|4"/1 (1V9&䃢8<||~%/b(\Wwxr*f,(U0xa23Ń2"۶@?ZoW>xeO녱5зQT\!EA~&©d|ڄT_AQX +&5*Cz Y=˞z@^sXe;ItN*kXxzKT ! |n[B˧ 44]ts̖X"JIKIZF 42=v\Rʬ̽Fe.MA%n7+0-h̯;9- -AƒEmCCg  {'#SZd%}H9DŽRnգȹMB Ȣ`;X+\$JGdbqruM+ru~_%g h#_ )h:w0.I_t˾ V? \tugWgf<2n&7E#~`r}Zq9[]D' ZGK!zm#Y_muLzf0@ݽf:#Kd}ER%)Q0@X,ݯޡqV{tHb:ҩJj1)׻I+9NrcxחeR jg+%><笁iqit)2;NylA+@ dz8IowQPc' .ffF&6e;BVz2~|c/n65Oie-0O4LU*T&d %4Lc$Nj_ GJlj#8 ;x&ztMoPh'(thPLٱ'70o]CU&`%Б`~>F]`[p0눲*X_1K&zGY; XF*"o}k*kO3q %JΉQ7I^VzVfWZ$O&^q0i^\] S/hFO˻ӍsyX#4ܙiRקICt" xW}sgxyc0傇&`>(?ϋG3XW>u)r't>~9IrNa6Kf}B u$(? od)peG&@^]z=H˟/wsˡ"6K5pP-Ra.c@K;~XUn:3ENIH" `!f2ClrT %>NSe J(^Yu`^#0+D`X4x5Qv!cXs‭iU"bK(F[ :aFhNPƎR;[N_>EE{*Yk ZyO'C(yFnεcBJ:W_UbzqKҽש~a5wc 昦9HʍL|o2ꚞv;#|MN|a5LG _I}ΒOQ)p qMĖMR|\h0/tv1ZݑTz+ a'ƲWsw{%҂4#BfhCHk;s6ELsq^O۾0+δ}'N\CNQ4M 4Aƺ&wH}u}䏕Z>d M|l)rɳiVK'SUr'S|K^)S"h~`ױq2|wׯ./F`P]?/ݿ6iK.0Z5Kt5׊iv),O;`yG|*m 1Z̓*R'!Š(omPQQኩqK-SaE ~00]zz/SzŖYUu("L}Sf9-dW, X zAsC^zo}@ Q^hY\=E9 ɗkl(1 d:d 14m#"9*=%}l@X-PQDhiqV(K5Q ' r,зay8آg8&wlعmTzغۇP(SFǶ(D68#|؜Ng c>&SC0dBP%\@jv@|Me?́=Jj=* Yw9G`H\Suѥ&HhJbߊ"Nn`On\i}V[7nVM7/RDio~_^]Yۮaо*Ɍ/Œ2Qi2K>}./Q1 NYK3€@j}tPry9+a*nr{*.'tu&,ykVx -jAh9 NF#?WFhrsqqfKɄ_4t޿v0K?aV`~"!5aR *Ljd<)V2""&ZH0<)c"ҭ\F\W"Ϩ"ves f30%0}DڬtB\y884[K% /馧mSK d{O ߩ OO3ݘ Xg@"-}Ch͊iVitt=en̞0{<aY=Չ ''&cG*jPB$QkQDjj^ LbI.x\r)A6K\{.wN\nsMt[id#׻ouvn}oMbL':;kfXK(G~}ˊJqR{v6suv:GSIbef9KJYv8pQ{c2o9c hfI<-/q[U""Nzk_< Û킧;P^5oG9N:~tyXxx.V5sX=6^o_=SH>rD JQ]`%M]\-K]XrU2rHʤ'4{p ZGE"Gt@x@YiOt9^s\31`E$5a3>ܬ#LڪI+:m>0B]Qytxua=Qb Gi:MG:0!Sc*R6,< R5,tΖ08~g ή? ]5 N5Wj/۝r_W<}z{깇,+3)m4Gi Jp*P`07 eHrh/v$!AJk56A#ˊ6é;yY'?_W,ͦlx[_X2DU`8=GknJ* Ov2*'ɕ.|7|~;b i(•د^Uh\Z_(~ Z.̿tёCd >|y-rz ,w}]vS=jb4bߧ7׫H 0, ؗ~< D,xgDmpqI~kXV377_Ic@?$:4$oI 7cI)kTvK8 Q9~}u{wU(*FA6pMk+_vϓ.T 5Uf(-7B %vE |li֕-@]'ATRrA_ )U] X& Ж9vIֈg|mX2+|d T*D- e _mHk6XQVI׊y|:{^eqY5㶚kMB[E!mAZwG۶҆T1:%fboS[K's[kFGEn~jpA (5yJ:ՓP0vY4a#A2sVk"ShecXOl9 13Q"$픂>qϪuFA>u^./Vzvy/]±׎N)׎w }fFFcKn"#Jьd3k,PV8d gSuuĢu$:~$;y]֜mЊ{SH@XC'ܑ t /|~ƃ{g`>侗`p98@_ƙ־g]yM{'6^;'(z6iOz%J|JDo1]„<]G3-&]{KeX$YzlSvK۩ԃLސytf PaU芛(-XQ*T\m44@ 3 <;SHC3KFDW8A`qx4Q Fd!FhPMy20|LqG+z쎃8SE:,7<`:Şoڕ=#~ث8Jʃ$):?Pσ:KaMBD.voإr7y|PMꐧ=M3Z+Z슥A7iğ WZ+-WråS5J$z17\!A"J t4w v6#&J.>ʊSإ,:uv?2<Y,f!Ynu '1beP_e\q"tK$*8XR1qV{t`E9EVF_{xUq/> ^&|xR/ˤ,3=V[ylUS-.-E>͂1Bl\BNk[-#LۭLY ˔ޠg+&:b?O^ę|.9&!PN3b&ӑ >&u{>)!+_*iC"6B,dM:9[XqUOlRp]6ܩ%v{HgSYhw4))~?|/ND_|W[~n+4o "%|rXبc Qjw!cL̆jUVJ!)Sbl:-&b%(eCh DV0GCM VsZcLuv{6G#aPzje*}hUkP!t.A~4Kc$H]~'»lmpa%2$6Ix#% arwzuջd~zw=a#>aÓTA9)Kjרjf5±&`&RGZJNz4"(3F ld˅(̯3I5y[Ukivmǟw4(P yS]J pRHadY)SaT:"HG !u*qȦڜXL hIYWtr Ankc=J|ט[?׮ ʗd*ĠYU.+M =^'jюum; c&Tx'I.KsH$s8$Zj9Sr E:c%tC«6n8u9p_@(h8\-ݱq4j5zG|xy&z5REA0ɂ U\\@ VHUSRĽSG_Z`uڽخWپL&GVe{~C0rh_'|q I*EHM #U쉲o4Q3 zF.R,GIuHZ[)5`<ѐU=ܕK<>yJ$1UWSbMLSd1ϺȮݰ؛_AT2WSXQο=Ksn8R·o1d |9iۨНoi #-#NDWW#X=hwX)(mw(wQOyxRNjy+XźIdO5>_5Ej7ˎ7g+uq>˝۹f#n n5[ NߙiVׁ8ϫ͛5ǧ鋳:=ea) ]2ؤ,k `F~uV@bvڠ򄔜>ǃQ^Gy8y4Y*QMIP{FiVɭX ʚBQZ-u]xَڭ.nn}ei'?'/_ @M4)i۾`r:,-JP ؚK1; a)VѺG,aElc0Q}|2BLTdfɶ@=KX+XJJԂ`T)Ctm@JbU3NAnO؀Ѳ###mH;j#a"69x{ Ҟ< ^z=[mz[VmS^Eir^c{[Vmz[Vmz[V5K{(L,|h@qyq>#藁*OEE 0Dp"{#!b/_G/)̓Vf[v* UhY/;{GPisrdeg.RI2LUquKA*-^@F$`Q;5bHԓl#l3c?{eۣUl̘k:fMEh0A!h֢Fc,Ib)CGf)LTIZk :(K?m1,$P(]#֬yؚ{E99 ֒ 5P8pz%Vy5[ήs!zsSbv_T6ӟ˙skc/DtT!,b*9hTcHLm0wޥ{/W3/:m>OR FRsnI4ȷ]ٓ~×]8*Lpphx6r?*M>*J[/*ëM`.-Sgj #}Z QTY0Q5K\.0g8Β+Etzi@ 0kehBhi}^4_q`jq3k1g] lŭx[uUtk{L-7ikD{6,17L?~8em.Y{H~ ɿI瓋Pb/,?B)D1 XpZzco>u¨㋸scP۲/&r!.j۳ӃWJVCRxmFJcFògCNZ<$ZՐ}~gZϓ1TgeN91֣95%S52>zI& kJ9n6"(3N'C$]0.$p%+ʼѡ:`QÅ;`,r(;:k&{0NfWt̚Qry_"|i(]`y{w?~Ćr۸+5b2TxӼO,bVg֐"F &&[cTaTHXS$,Vh0ɇ )2^1X,ؐU)H29[A:R) SĠp)679j5OƓM^y>v+Haaơ|A! uX"#:5f\I:y횧W$S!w-5.B")>âɸ3}掱C˜ U;@Il$\%I\4uN\CQ&E-P%s:Ak萴լA '3=t2},/ z KI\Rc`廊+e {W./mFѻG\WOml׳>9/őe}]ڏe7 ;U$R$yf V bXe/Fˇ`^>/e hHB6%!ime GCVm539j{ j^|&?.^O=^w"76kuLq y6'?{d\7X]hobuNgLz=fs>4m~ yFlOޛfÓ/C ymf~h[l'Y6#f Mw#dz7jzwL﷘RvhF?튩Zvը\ٮѠv<}xĢh"* >%#1>xe=DgW>%s<>fNy )/YE-65N>IXd>|U8>4bq##rv1ψSZsj Uڕv{m?.Iě5'g8K'z}ј&]T\M2V寔LD(E%[JjiK)0%*M||n#H6)ˠ!ϨOT*HVx&ydOh< 11>c.:|'ySzWjJ޳S5J3Jn]Ր4BVUDbe1Nrutm[0_7PzqsK,/{/899]=`o*0M@IIc]b9UUC(lͥ6o{}Eh]ɣJ0Uk61l5Y 1QEbaHccc+)P QZ 9е)Q-V9S|,jp&:cfَ]"PZpF# c□ N{P&td^YY#VNfqJ/8cR:w'Y/a-'dZ<9ybZڅl. -b=s_TiHx'O2>de+  OϿ[uky"(w,4~Z{mȡ)1-H!콎Xa^{=`}^>ڀr?!<]|D.ݧI3vAM5x-Thʋk*ff:j1ϧGF[&,/G^(Zj>/v)% ՒAPq#+|J.HewA?mh$GuIJdQփq0$9fT(%G( i9sL(QmGs+:2,"7 -J))$ky>8\}﹝[\Ld_:ϯ67x8:u3ۂdAs %b`u9 ]N#QKzHTjc!ֱ, RW@0ڕJ2q)*Q+{eu DkWg\`u8\s%BcՎZ RR_N2y)j:Qj:QiVӯPM+E2˾OIXeQg(weZQId$G}?fn{pz0m8R ?U}%]&M/Kp3"ا9KI QDV,s$#]M2Y@nGIq ~xf|nnh{)J(Lje>弫4edn,Y0'k;,-ru}fBB2W S Kj0;ɔ>G^&ky`'p׸;Y.t\f[ hr}rk!Y]q:ѺUoNȗF/;vˋޝ|,q?:ctF6fIjV0wh osǔe2٤ANqh<޼ӭd]x':j܆A[A9MyQ|p:(pAleytPLA`Gn5g %0GFtZ~2sfrvKLpvapd!{ڛɇǽoωAܤL; hsX :w2@nݦk;T*4:ʶ~,[Ol0]+p#:q)-ٔ}$,S3w_P>"W^p(mCxMʺ 'k$@E]P,8 DԗRaJ^a۩0;."}{L_@nWWy^2ϣobu%:tif_r%?^]%*+0SL RW@0b+u*QDVWP]qA DH N\)" R6}i-~4+]K'* ZjiIX2Qkybf:\gjZ!h0)p:h%7iQh 5(O+L;ڮcwu&:3ɳ^y/g%X9A[fF[JrS ârv a+{aC1'¨t 4UvUQ^05\վm}~44jIH߆uF/#v͏МltsK&k5cWIM~@ @/t.1'OVXq2DJx 8+ .: a@Na(GxMUAօr\`3 ~}detPnu"%|`0Ie%#]4^`rXx<(:t[u+¥&LIxqZmp:K@nǹ՝cӷ%0tx>R'Q2p% /*BUοn/6䯴>}#-˲rGwqy64%8K{)y # r\ {)}LƓgd\x>LPu;؛"%1d^afXTZI%HD0 A$@2|uC/W+ƽ P5WΟMXu0ƒ\C @XlQM*R:1ga9MK] 3G2\q XH^zX YVDykR"Ĥ2 RtHZgk%\X!0Jg戮ϛ* SI ?Z餜naK@!LޮyeY,g5l{wuiW6c)dFOo_e1W i(q/rԪF,?8 D?ſOXe0[Q-@,dx77uw&|[M^̣]7ʽY]$ F%sԩxҏava&3$>$_/=-8ǬrGpX@v$߃H*HʘY&h? 5*'zzRTɍ$M=ݯ.އ/{:6`kk<'ouy⒏"4cDiqWpѲ|>@vWԴ>6:҉O%ş.)gyk݁6-0֯~E{ՍH_Fll_^g+2 ضT*D*eu5pZeFλŊ۶rL:狰B?;$^m݆J~ubqFӢyR~ݹvƧ/]q.2K`y4H)a@o=/ w{06)5{K)Nb縷^h΃/!eP4p} B)EX6X~JG."FUA9⣰Qh}>䜯W:㉃z'&Xu;iq 1}M7;3]JXz32<]$|5N &$HQ$jDF1u'lJa^iGjdL10ѰKRUc?3rv [Oǿ`#[S4S񗃬#g},CPԘr5nah=Wǜns >6"f6OL J;Z^P3y^Z;`^//o?r ;/`~ |_}r>]z)!}0;EղS3vya<Ɔ ^T+: ~*ƣкO0UYݘc\}s{GdmUC1~g6^d~~2F'TKhìû"ez&c~([ F4bz•n_z7 Œ9K˭.S2VSWU_,Y{a3laU)EԒ)E 0ÏZd(xWbeLtX\2Eefw1y!Ptx~6 a&%]A^{׃li%% l}|yw/e-]DT$D|\ɺ[r`td(^_’r㇡l)ST| b_ iUQ^ /tpmkhirxd1yQAqp)Y ӘSӎ_67̦J!nZrjZ<(^bWv:ۼ9F .(^Ls"KvN*lgU 9'h)p!)Gevqo㈫.)]± IUpcK!zƌƁZ 1.RN)4|uW+}`u6id-%{?g}^%}}_ޚ׶vrXֺDGX 4)=)~\o5ڙb2C=(,ph%kp +ɘ`X,<|ްaݟ8V&M ɗ֙.R9&!PNK.bԑ>&CvbOl ltmqmq/mc $K*{eׁy\`X4x5>bhX)ceqc*?{֛H ]dw/-JGJU^m40v_V7(0.(0UvmHȬfD|id"2q[.XK˭4$Kn:#DJ/:/D@l­;a=^~5^|?Mqk+?u`];1O.$߂sM;?Ϧf8u GZ9"&Ѽrciɍ=rci/X,H9FR9MrO 'əV!c0]OiWt^J}ږ4]ixq%_b` ƶxZ8j@] @`j0] \ɆJ{Lֵ t@8I/xt+EJ1(`b b ʸwZÔb{vy#2re"2mqG6N8TX4 ,(&$T6F k&&8a.x!GBDb,DR9*;NHtFN˔J74 .ro0megSk!x~ 7&>~F؃Wsp6pgjN{K2$VqwzȀ?>吴F#]?F.L?5YM#D5v>q`[#C 6Ϸ~wqmxLD<L\'v`J.1pkzb߭ /ǼfRW<%Iݕvr[[9JQ$L 6BF; b?)?rk>ꖓw$%w`₏0&!O0Hu´$:9gu"zL`SIoliEbE'tXs)h@IE]"0̉l2߷lo] d[ wxÛە{Jx[ۀ|QȣfAk2<VVEM$LJHIN&kTVt MRH(ŁqO3΂:D!Ra .(CR08e3rZVE$c$@rE>+/Nc,S"ג<.ȹO&fI"g޴NMx mc=A8:ǁil@``@\P@ZCdA"yEz3H`u}uu+e8j%:JJuu꣧a2F]erj }WWJJzQ: u l0*PP/2 ue%LpL.CQWKÏ2Q]aRW@0r0*+PUV˾LE]KuE ~C8AbQW@Tt( U\]]ޣboo"W$L u)"Ъ[+F[%3'A;J<)rߧ| +5 ţvOk?jm-= Q=L.SVUa)?;R6H,q2DK, @Ͻ\Ibxf#8<12ɉXC $*۠7=Fԝʭd_ }/pW_WLc`|oY{=>Y[#߽P[5,ʫyFX Hi/zbe ~/2Vd@I+V$@̑X[G[茖] ~/z@_ ^ ~/z@wpHTv~~(aloc~Fn= $9ɕI1HwBcSYg#1_'t٥U^dͧNiR9xx.`]ChY˴&c6DL,IE)1AbXjpDH4kSewKes݇mj/PxGY!7MG0.̯`B*?FO? \iAozgn::m(ϴ >~;.̺K9o qߟ#dTZw_[5[/j6ӺE r?T-4ٷC5.|ڄJFv4_Gil@|I\\uqy~#-saM7g-LcJ kv(P%stOٺx5'] C&*ʧ]Y^`ϣx>jv `Rb F1Y:</Es'=6 gP"YQ>{9߂> `Yc ˞k ǚxX4R(ךZ V!_!?=X?6 g"ʩ9ejuqTnA?GP[epne?( >`"-JąF'0z-B]d8](=i.*E:*=1"#WV)1I."(qJlM5LïZR2i5LQ8U8MzzT͙Ok3{` G"XgGγ O_>Ӏ`4hz1;&c7 a$^g3?>kmM]pek-_2̫S&rfδTqxLhM,(8P#4Z"K by^[d2R%Ds'Gpi>L[7ξ/ LLl"7Enz)7DĤp !rM 1A)FRsirDi#N+qUcB׼+ȹgc&6)OGw~pᕮ`?"ͬ^Ab]߰A(Zl&1}SF]\I|~o٪hU2֙.}};itAMyQ˒qvtp;˗ˎ_N7QsCDI|`3\jCkhm˃%lSc}k#SbW&E!ΎBQ3c͌g^r@oRը`d_+[onR8J5{G}f%ؗ@|2]`S*Aj!gE8Nz).۸h{]Q+BGŰ$E.(e_IԸo$w%e_+9z Wn2~ W62\~LT~wx?$ҵ}̶b!$0)W@rw^>fsxOcN1O v)NC ˀ;sl5ٜxƁ S3R$IRp O6מ0{ƍ$O nwwU pw/83(YM=lʢlyLzu=erZ:u0H#ܥą`P!xc|xb!&ࣶ]N`FsI%c`Ipθt+/uJRbA,$"i]ZTz Ӆ*?ʁz.N T .ҁJna6eFs"#BH djwZ/;k"ĽJ7inO:S>诋A!R m`Z}0{3֬7ȕZ9O ^J4CHFTHgFa,R>*aɘVY-!=%8=\V )Of~>-v*EMms[\Ϝٴ]_ѫ\#͡<>[f{[5vr׶'~;J5ѿ1(F|{3RKa3`5YZ5_+B҆//ӗ;ZբӈtGB\^o34${E< W>>1ayGdg$~kPV⮮HϚǬcdD`Z $dJI$$Q)D1(}XCXˋkᏕ5%Mi(; %bܸ7ysO:KۻaGeѝ^%ebs,`rJYqP3p*ض|IhiԹvC Z{8_$Cۀm 0T]]rBA\]њڶZLq+f:z6_pt֫Ǹu*;,.?#.iS-`٨yR~Xn ԞƇ Kp6tDB2 ! VHr~f})is(G>hт孲2& ޡ/mҊH>ɸ=O2IؔXmH66{c-iv61lRJ&d?_'~^%wOJC21~99޾wz[nt}D"pOkShVfqZGjWIK[]ꪃX7x8 IB Pxfp$7 W:|,jiM")D'RYۈ1M%f-J8c=3<1]Y22; =bnܯw\h+UvdV5˶LTz>f%tzv7zQMp0.R%)J+d*!Tt/_A;Rg_ξ*( /(4WQjx4^D0Z "rÒJrX:(ng aNȍ t@4'#nU{2:#:fC,Ec`g=ğ]ׇ>FЕەXڂٶ/$?)[RJj \]Ub}4*2m\_{UM5bPףK'7W󡋊P#T'[W2ʋ S1[{$5V[maX#5VBe dޛm$Jx" ِ9(Ҿ]ǜ2hNc=Zspc>H+vT_vw}Etc+^3:y>f0ݫ*ǻTAY2KPN8o3Q)8m3)L&4JKsY*rΪd9W֖Ā~28lt&tqA>氙l  .'.-}#'Z-g,\1,Ot$'hA/]_ٲ<0RRBĀvE䤞VY$Q!XY- 'p|1/(Gɋ-L|Ѻ= -ŷ<+Q-*3.dEk*JD* j{F| `iT%p8JZJJ C%7XWK8*]p)ZmUo]iУTދGV;0T,D\LAAJq&4/YC-B5Q+G]6j@ X̡NYƔB.f1$nW%sK[Uy)=]>K|8_9/E)zbL1 ! L:ۤP*ciX\q߷Kmtmަ xWKu[bȏ跃4c#9Y `)z\5x.VqwsL`y=Zᱫ6ѺlPj*^~Xa5k!ٍ2pVi4͇FZMG1~){MT}{4,WRΦ V &h`tR_ l@+y"vu I8 ܗ]8MLPlv5;l=G ħ:7?} /eΟƴ_spި4y 7&D呧 O`1i+>GͲtѝ]tH@M^YHAiY CΩ笥BX?|XWsaZ  ~iA{FVφ]plUV^S-Pj37ɮ8A$~/ъA,xfV,:*bOb^UuDZ~쫅w^CJRo2,7^o2,W%0TɃcд ߫la q]Pb J h8JuoN{*v"P(bz!s"Gnt i$DQfiKIĆoM)zcp[,F3'˞`r)8p3pq8ͯ.׺FKs՛ʓZ=-˷_#bc2[ZK=&\OZ*-kEl쎩)ԓۻzNҠgsI{{Y^$n]QO]f`f/UhtE/rvL=}|f=!hd`n[3=ycc救)loqw{w'7yWwLTH8^+Gz_ ݷ_uK]Ǽd>x?T|*sP\9 gAJq;v{d'g8sd喸 M?:+<[HDrIvx>[(Y8y}?8ف=ٳr2l uqkr:',$OUa ֝/X咒 V|}s}͉&+Pt `>EJ!:p0Y$L0B-A}N͉ ["6$ %/ɉCBOtH02z ;g7iP馩V|O"xg?bE=vg) |5]PXJHG i"XP4@Z܎.8,Щ#i纠&{E4f gLi$u4In&>L*)!Hyi: k16P@iǤ#Etga?mQC8iI4(dj-M Wr{.5 `^@)P %ΣjY]{:Ӵj&h^qٙkRYRC0B_`ďGԹ= xAbxAbx oy]'IG,r$l֊Oy 92`YOh">r0W(Meu9 Ú*XK,h91%btZN *).E;vY ۛ3BW+@rM#m¼{pnq{̔% ZNO,i!&L1K.'T`N'3]  "D)D5Ϋ S>4Bj]J\576wD@&VࣶHD+} ^1 0+u i8fI_PP-XhJIgLk$S!1':yХt&V貌; _P>jĨf_o@c}1Yhb,UǃR{Rdi@HZK" 1d88>XhlBqU󳯚phӶc1h^{8BI盕ezĭ_WQq#* >W [ Ds`eΕ,t7GOxruUbM[SHM'_db:3{G7>Ѭ|,!癐:,r/2\ ĚL@QfL1 t*sap(3U3C2enY,g|7 ёn1n.|Ti=}.z.qF~5/4覗t5fP!F)ciFXn $%]x=brDtS@ B߹h/{Vv~y2դE#vlvkn>j3?tYe>Bay:8YBW'&(^(:0MԸ.ZPz_Յ)tPԮkEHv M7NA70iE7GF<36%mYa#ЈU;np7] sCbfAj ,rZh{(^IAT~4=/`RYcќr6s/X S\X}塓8Z9(FUxb8j=G4F%RQQr"`!_)Pʍ؈[֑aYqGkZ )$opW>ӡ/9dĨyjvАŇot. qi2qӭVnX!rϕeôrُ>.(بls!u<|J1higH .a!H*+gyÀD̃2éMWmj>?n& =n&kZȲ#͚@,y)y ihRK d 5܌hPHw"%1d^afXT~9_lg i(z~55V ZN12k1oͿOÂ"٫?s(7P(5Ij e>Y</ e|LFW`?U#XtFd/s_!ag.{c&ބ8ݝKwm;sq?=f8"$*Ibҕ|NH:%gͪEVwa 4^ߜ7ʵJ9eH]u} ݇cgU -1~-xV.%"4cDiM*- dGJ4݀T wd$~俦}.uZEa,g~ͥsvw%ֈmBko{Y,CQJh_Ak]S](ik-z2i/o 6<^&nqRYޡ` 01hUOJo;/@xwߵY. x AF=iR@^L`a҆T1:%fbo)ItEZh>2g 'PWiZI ߎ)5?j'4)E E,JUZ @ !(G|6j?wTUVU$=Uql.}OȂ 1)0)m QE0`ţ9^9 2x+ z+(9F㕿3pg6d揜/se 7DCJ.e3sm0un2y:Е)$T"1vA;F<,3ಠ2s$(k"J MHuDuŔVr׫[5*Cz x<OcQ*c٭m8-/)#qqh=]R{7Vnv]5+o]Q89s\InnLs3u GH%7ih 5(o+elzDTƅ7LM^Af[ `Lxp]/L.|^}Bې4 xP2[.Ϡ}(BgYx7,rlKڼ?Zbivmi_z6mԵNIWM뎾00jzzRp?|:VfTi ڞ9Nn}D4`?pmI 9Zr;Af = Cq W? [ePPkTÌ.5'Hu HT D9_niJ+$T΂#"UXxE8n&570EbFHC(&P< oS~Ni=6u_?tlhJqobf(I+o 4@=2 (. }qs`jsV+z{"l KtZ/Og=7o4tɡ\JΫtA1ben({tPFl2-/r@PsaFLIDp*wma޼~4t< 3*˼\e1 ^\3tny$XLjA+oe5uJ!~n`k@_oB!gh՞2#`Hb:yTk%iV{vp6dnf<_X"ղ51m Tx`rwn|}=HW xrfT? ћU*QJc"' .fF\r YiTftcH=jY^$Q,e$̋BԞ*L^Tܫ"XkB.0p3k Cm1_:_"ۀ?!Ҳ&iʩf&@QGz}VU{{\x3GQBX\*96Ns0nɽMٝ:TU z? raUf}Uocmw;a1}qT-poL^Qrk;3e+ā3~vlxZVϷ wկ?Wmu,Je @ #UjyH_^~t&qyn#Xm]5[wqJuO)Mm:IBP:TQɼ67'L;L ֨:.nz<,fEll6k5*]XTc']YdM9RR+7Z1O{-&zGy3ǁA/^`.QۓiO܅9R՟Z1Vc)DdQT g77:QU3HIV]^C*,Ov6Pɚ.㿚msʔ=8SC5%\aBSޛP׳\mxG/}-oڀzOTv08u-83}0;{h$?6p7:8^8m)y~1LeuvLV:@-d?owYaY>0vm|EOAjO0`D_a},XBLD&iU G|=/vh)8+!PNsb&ב̦]`N)8>HO td'ob{9v˱_ T$\RI+Kkd KǒFIOfQi%hX)8`1waʽ3ø)aކӡOZ(#oy`}, ,n||*LbEylmk+њӲ`ʔ&9e(k1Sn8"tNetN;pQj~vրZݽ6\zP^bzGI a͉ :gV\I+sK;y{q{n?&MFU!KHHYpQ*c,xu& S"`"RSF%0>H0<m8SXRn䗯}8\}qwdwU+O* J:x0x3E WK|`&7%f踂d`09'Z<*u%$'H] )QWB*~ :Q]a[fe<ׯ#x|=/?׋AŜmO1EX_@ TJ VmDHgcf 4s5)d`R5˘2@ V, 4jlS$f, G\'0:Y$'AOTi?d])ttr[FxSofz%ˉfX O9LI s%Lnbv8 9Cmkss EG53Rsמm̔귻F?ȶ3ڣ2F[a<{mɞex9 w gk*0#ǔjhK3lyr!xA;Ɣ@!e3( | h5c$ߍ&q*ͺN4G j$(Q*2w!إ)a $ghm Bckcb $l MK #yFugr=k-шGj=l|;9_tݘȆZlPѮ!}"{2:O @Pŵv,~/H1'^D;56CCZ-C?lP& $~ vW!pkYRT?-v V2IwHPc9\A29ޏܡ cz V R uWv@ e2a?oЮ CH8PeD&Tt "͓l4#Qh4[SjUԭ:A1 04ƎIy@C\i4[˒PI!0% DslH(u3Dfy!ܢ+C*H2Y1PHq s$e七`Q VmY2 ΈRl eBREA·]u+R1I+xs JHoGȖY :pa0L!CА]f%K*CҵLŃZ2C6a1%OHv9 V".b$) >@ E&rZ'd^0P>TL%F23P&\W]dMx+ q'2Cp䁗>VVWUSYߙ%72b.AہjFF#AtDa.CJvzwF]BlÔ6fi2}2 ][{ #C+|ui? yj"!&j &: 8gk ) a:2ݏEx8gWB̥ܶ%[1FcGX '$0ZKC  {,hK9CA:G~"h-J5^HݙPȤ td!nda6*I@B4#5\ X.ëCuqB&"\G 7åUeXJNS NJp"*iBPʉָBŸ?VfHWmY4MS j$,Y+" (m@;Md魛ޫUߞ`!-3!ZHn|ã>&PImJ2{@et{:M R `-|lыmdF؄. }lH(Ŵfx0Ь9Yk-)[7[Aq^,R, Ę +@ysQ7{#dз]Ufq(ٯ:(Z7d*# %t +t9MN6a>6[7I1c `V$-aK}ZB<$܈ 9N)I Qރ\:jD sR5 =Au#䈁g=:H5-иg.T:WМpƞ #:d2Z5AaMJ%+gWV?r T ބPmTcD=ܲ[uӉx{fCAPcv գUsu@ D k (8:z7%àeE(;̀|1 =peq.dvčtX`*2Uk57eR q՘-Uun4eXN6dj4&ńCH*8ėԅ˯p@BpC9TIVW<"PsLnlz؋[n4Cp[@0~KNYpbc(NYEx3,a(/ %#ߟ!ۛ-;P($tGx;mxKf*/3^1Sn܎h@XT/Ls'Ř: N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@/ $˘nKN W .'T' 9NH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': u!Sޒ 5v@b6b=y'8I@/ 䡐MT': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u:|1!'GCq \ޜHP:P6R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N u'Ї[mZgԻcs݆nc}g@ : p̛1. db\Zn\{f05.K?b[qMG ] \[+A (Ջ+mqCt[ \K[+A{pAINUNorB-7CWx3gt7ߢtuz2>3]3ІB?}yt叠+t]/beo9w{% |?{'_wa웳/VBv Ox~}Gg}/wv:=av17ֽQ.v^bfg׫vc,K]bw5vYhg?gW߿zX|j7W7oGn/SϮ1~x@1lG&] g~wv۷nY}|-\ݍyb_ۿW%8OxB%%>9zZ~0((5竫g_]ennYս~>+=g.=ƥju.Av-%_o)Vw*7E~O&k]5C7bMBֆz3LcCJIn6$\Hƥ2K~+LF>uA&(SPA9|]#75{t%pnm:QPUJW/BtBWNeJW/ ѕЕNWۡtuGv=6F\3NWG<3]?bqh3}8 IAWt]oe6DW8p8.:] J2JW/^rb{b͙oLlMNvIp~mc rv"Io/CV)%&$cc|m̤Q'E^"9m8mmLNW2)]DrY] ༝rf3t%h=u{^]1q8-vS2n֞P蕮^ ]IFvKxCz7JP?|t%Z&!`҂6|1((*]f*[xexÄOdˣ4흒xקw~W}%fTB u9krsNk]굎M8I/ i  $pۊɜ٩zȱwa䛁9o;9%(xߎ¹A~ PT_)'3\mxC5ڜǀ#+DZ].tzDo84Oi;wہQ.tꝹz?MXΦR:q@sw]{=`8IKzߖ۫ߏ>cj{_1=`q;`fܗ\ E˒#ɞx߯ݒZZ%{86.T, ay!'Uˌې˖W-jMQ`/j.&P*' $Ϫ<~16Fïڪ =\}ag58p7S&CG_VI/(Gwg62_ ƥr/1NOWˈ-,v"7r>g˳VXl٣ݘwCi ƒFBMG-AdX驆Ȝ7~ٟ y2q%nc,U%݅yyY<WaTn%A{4|eo *3>XsOw1BɽS_Ǔ/g`ѠߕWݯqϚߪOu٬T]#Fq(1U_,jOFw7MjŽ^ (ŀmv)C=!=x<AEw;,r1ce6*O}8@w)_^kZ2PeѬ1Q~Xn~mF׿~8Ȕ##`rPF|j2}\4_DԊ{qImn{es+&.Plj>]^\P?^Z- 'w)Y{ڻ!l ܬ Kس:{òJaq ǧJBLT_2z)/8ycH6)[K}P*&e /bA:mC~J:rň )F8CQ31CgJR-D4&hc1Di 6PJ3.Y >pOAqL3rk|F1AJ._uvnp+8ۛBE|5?e2$|t֚wDhN kp}ݒhvT` bzz"x3YWxƅ+4UuS7L/6śɸ<\wcx -W0³ cb==e ckCxN0egc2ML6B0VV1oK MlHT$NRJhH1)ey*#į$N!|dH,:#5@HcK4.gr_N5]vjmwpwNc5!q3)2Is&RA:O$3 + DH}(mJ9GX=ͫ~$r {LfYk9ݺ |y)pz@?fvOӵ>l7li`"w6[発e 굙}D2_Hnc,vEU{w8ٞFuv/q~h bk5}-7j懼*U8g5pqx'.7aEUoϖLPyE]y][ir |sƹG$ ~~He?ߎ˾oYb8^Lx "M~h/<*-bl02j /Xix-ซ< y19Ѣ"ϋOHw?|>?׿kہpgMR6$nXb 12d JJQ+"%ԃ4}Y['ۖ߷K ,:tO1rZyi^ҩfTF'=y18ezbG5$&%tEwJ*5cg׌QItagL]D^U-Zr%m8a}Phe4|G](SnxHav0QX9 7(4zbĄªb"A.|`˛vz=0Q.x#@wa25 DCo!Da$Tu!a—2dJuS4,H|0"ф@ ͝F蚳Xw!c- K1%MK݀ob9{ٵ+s0B):S2M'Qu-EuGYB!  ø/ ' Sqչ |z<-LSZ8v6iEG! ʹBw?!2)0@ʍ34JHwn2,8AI kUB% $A$ 8$#]FN9,fyv2=(Ȯ22%À=lYöoĎmѲC$Q;!h8c|"+|iG  ()(5Ϋ Q>piԚ)c8k{cs.ȤP\TG|N fL',">yrfH"$rϖ182vN2mbDARD$!Z;% jIy"$)w3i y\}ײ ~[:Q6Â;A}84|q2LǍ9b G +D𩰞قIxBS_8 ¨M> q#u)f K(Pag:؃򏅀>6O|ߎvo(M/g4-[@ 3tnzEfڢڥZ5159Gi^nz9KA%q\f9(J):S)"Βօd Ȧ;y,hM>ZϢ5'#k(z?SņY/#%%sd só%E.N,XPP\eu҇Z.N6zI6~g'#kۅ2]Wz>3"ZFtYxISTPI+(pDgVc̐ LBm>K{RI8s{`7gG&>ؐ2@e? 1 obޕ2ؤΛ=e~ēy[0֙.I[4l(^+5W'N:~ryYrm42eQ (Iz+eb6Qclt"|;j %uzh1!$\Av͌$3nuiSpݛ7<~srվ3?Myf?0Bx7_PCF#WÏX7]¡b_d5%' n_Fn#2j=./R8G()S#e+ΨK9+9Sܝz'3{ CDԂ>RmZW]^2A^wMluE;R7nկUQ%T_O>)XL`qxV)K?0\=$T{ybmurar7]z;ܣ^j~K5dggF8 5UW8U|!HP"E?t@FXfYkm$iMHEw IZ)LҒPTi坺\J8.p"+'N$J")!d`$PO1&$L+Mx6jbMC^jXR侶߿ZWHOl]:[G#6Tky\(cY)(vyHT"&I qh\Qu̓u@+ĉTF' >`wY|~Fɗyɗ.LZrȌ"{<fedW^䵕ZHFew^;mP7t"C,S0- ‰d,z蜣,PBw%@ O= *UiDZ g4AO-G0dREES|R[xJ>߆!$J"g|צAB`')!O,!tT#]]xv+bM ZbXo.n}? ?M[ٻK11oѻoڸ+vTQ.TvZwve[-,ouIO%o?JvOx]yw{0Yu'ævYhCAm-Eo|895-mKd/o28i)xk#QxD.#=(8o|rRkPmy!u(i`G[6,2/YC7k9s'6׏o]X3IGYO/uA(j XE1x9{HԂH{.w$*~,3c'7~hɼ„^g`' n??~܏ Wiлq8>[tx揄cO`uS`_YE>"u-Sq:['=-Z"'$?w&A(j!uTp) ^@YkZGËbWbUW^LiB9ʮ.]ހVV]sfv]aWAiyeW]YY2.]!R\*Jv)@YUT+J v` ^@kDJŮJUvu31ĮԔaW\/]h%:B(;KTvؕ6vA^pg.Ɓ -;LQW<t2Ûjʽ տFЈ͙# " ̕4odɥA|o&Jt ᷹;k|dA9ܘ m 7z0߄Y D;| pMt*l_h MSl=|,O'd{y"‘` w2YNVml9eI 2@-[t}u2G(԰v\Îkq ;a5츆װ3=ÎP`6 ejCP6 ejCZ6 kCPm(SԆ2Lm(SԆ2Lm(SԆ2Lm(SԆ2Lm(SԆ2Lm(SԆ2Lm(s6EP1Ԇ2Lm(SԆ2Lґ2T rp'ډ#O P'\K ueJp~msG>$ J&F#6@Khej)c]NBQcWI{}6gx("<Rg@"74y$ӐPSGNwOFB츍̘`3J 60s!>}F9pJNST"Ѷ/wHeO#rBZϊcvwor.Y-~vuk}@OJ3Q^IxG>QE&bP涍"Jǃ 'K@xdD%Pe$'( 5XE(pDfMAfM,:AX*ǝ*%9\$4(e0¸VYLC*:eYZ02K4\,e$LQLBA쀹[f,䉣aJ 0^\ qvG#ˠ2|nS٘g`YRx-Ř:\ד2l,TN1"R|^/N3M@v=/y“Ts%,IK7goHFOҾUEe،7"Z[%ؑ]R5)Jβ$t(e آ kKLv#1X7]o>9Mxloor,]%{l'|$Ro7n[Qp~"|n@;Mu-*De˫hl֊B!BrJO%g&2Ng8Dǎ42st-Nfk$U!Q'B)7\)"Q]r.,Oe2+[(PLWgiL52PTWcA=x)C2FJvBˈRK<$@61X(g_4d+^楯9L@%TTy,~)U t#BU_aGT-?Gn#~ɷhy h))]8u1 r @H.>Bytkُ+S?ö*gR  ?Gq 1Dʔ#9H@deRJa0̴~e;X~J0@;erfP 0$+p0"e悷*)>P )EL$3GyJ{ZvΎ o~5ː7E167ģXݸQowC~cg!kO6T>OIIF"?N'==wy_ _]鵝 ڸ/o|/AhSo B.2V`8fe4x\l<GIغc-'? ~SO^J ϛ]{.8?ݼɭOG wh >.hnU7?n']/јx~K *>H(x$m Ac TUȩTvq9*SnAvLx\퍐YYrx&X`ݜ++㴦zgA N9|YEOlX?=9%wsI?/GDv d(rCI YXN}6VBtAJdQNU x .zj|U)$lC,YぅFs!1kMMh|[fi@ *iڠGJc^Y(UPNYi&FA>H*)E2! A Qt-\۠ Ji*.N@-F^;^EEp30Mi*q0~lH.RU l!%*"T2S0QF9YW{p*tgXSnN`4!h451Jb,MV*FL֒2JLizT !]֥H^gmm2d yf1!s5Zi*rZiQ94Y]JZa*@1)u8[v釯 u"ZO U<)6use?1o%޵ lJ '|&sKY$ Z.F[顼`Bׅ!65.}YnB݄}F+m$GElgG*ayY` JRK7Öd.T9 (W9JEqCoZ~K3VT A4!emD$pZx[E#g ZɮϠv67NoHYnڪT١⿰}62u|M+lMrw _$ŻJ3nj֢ \ST,uZkw||9;|3;rYwZW7f)^)vy<1˛S͖>j)G>:Ra?fi^zס\Pn!iG"23R5wL̔vݙIX.= 7G7\oI˒Ş?3L2c&^3AU= .a4cKTtMX8\z{YQi.D4!tשrVgheo (?5<cކ ' F_KNKG*8NxLIeXֺ!,F<&IOTjloZmMZbS|wPbퟆ^w ZumGBWӑݵt.:DfH;R+I""`WW\y5pUզpUt7WڀJ%zȕ`Zxከɷ\2ƁfWWDjuF] \ع}9\\u[+kyEp5or=U!zPlከ\vpY#""  Wv9p{8"X(qa:W^]G%o\#tpu[<]X/(wo3gy(Cy e`}d?뷟MFP,?*N,z>p%@f>c4씼LW yL"aP 7ӂKyq͗7C oq&-yiAws:z ӓB@h:*w)W[oD.3K?hru= vo7JhXY ֨RwA 2LKc6evZ " 9dT=K[9eHK̝9?TYbQxq}wp mI69.CK## spcg61;',הyw5vybч^2Vwe#3V,(qxr,aB :<;iusfMSߖt?oh(4h˳Tr?CT $ave$/ 4g(Ր(uk8o߬y)wQPgmM^ I cZ+!P>IÐ NlWjRRYE~&<_`Up~tn~=:Sqz=it%φPل3d].iStBG3fD%Ҧ*v.wgf6v5uׇ^)7 ޫlH3m\H!5ٳ7 4ZեeueIЃAm Jr*,D0vΏa3Xk=nMD@hmF$CD<3DYW=JyvXW"o]m||]?ѱrڸKreNNԻg͋zs%Og7󵉾SkB"%dU02VtC(R'̝);AAVAlԂQid ]t[.-KAh $똶cgӨMe/rk#)6r]69yRwvJZm& obi2f_ۻ0,btW6M_r1XfׅN^ꄲ⊒.Iyj &]ߖ']˶']*AtIo0B1fmbh0HjlWQڙR@(BJ &kI6A̶G'=z%׬s]6@?\JmI`Le-C,(-YРl$ЬYi EF:::\G_YƔ4YJ o@SWj<5mL՞Lv2Jiau+O'+ /}HP,$&HV631I\[$pBKHB\s(0$EE}J\ `]_*iLB9CbC@xF9Sp ! $0̬cPRtty)Q輌 @KǎqSHؤ/_ۡtZAh*vڱV1>rnR S"$mK,$$Tq" 9S{]TWQǤao=&* (3SQHRK"$@.qyc-I% y )Z7EU;A2 Eb0űYŖCjw7?T0FzGT7^]y.аmU؊x+ֱ~vօ$-5_lwȝ&q+P`4J2PG2™h8s :t־ N`&EA%i=ä BhnNZ*2BF}O&%[f.% eCp q3@4y B7{jw lB 'Ejd~H%w[elnMWVS`|>OZ{ͅgqyy>ѣl̰ĸ={i#3/G{"\dFW4zet4鞇>>ݳ9 hdܭ?v>y_F䵑CPvw>dϏ0c{l,U|ko%5Wz롦,Vc^3C}KuXc>k[_H7wkjQۡEֵpj9ʵީ]Z!~ŝLFy=}^0>#ё:!I%$[|oOWz_Pl_?Ǻ1^^3Vpy}^\/6iӚ?=͚sO>pr{ha %*f :#R襇*Eg4tD  -/ bf\SdV/E1A 1Bys>c {8#g?ɼH9r';jНW'A珟\''Br]}[iyqFL4>s&&)r] *H:P{J]{ڃPquȨɈ9Vjkk1K.zR, f6gJsg1r)\ 'bfP ϊ>`myp}gi/T~0}_+#s$ F[[S)8aH\3.y}a}0B@p:HfID15 4 E($D&C@sYN26؍hke[Pq*jQ[wIM@UV@+ mH2L,jA]$)=O|Ew cU:xli% hH-(0D I'͍L6FR )M,jYYYDFg4,(U:qmN6kMU ),+L6e:C* yZ_aa27޹1rB)^;U~m.uN֏W6Nļ4_kє92\K{>_#1664ƣ֕VTC\Jfλ&o5 ^ѡԅ?ʯ>✷&j<}>g'ˇ&Xg20`Mˏ>~uƫw܊x"Hf602, |h3N "y+td\c[u6C[a9`Jx;~|߿!ٻ߶$4~1s=fA?mɑl'a&)YE(@l6U]UʹEs1ڥ^_(o},M^ ,R|Jc@"k/%%DS\Y,-,xGI4 Gೀ r+!kxUfޔ;300S#P--X˭;v>fru_zY7<2IYxn> v[RI=CKY;N҂NDtadk_WMˍrUYQӼp<(pAl eqtPLA`Gn5g %\0GFt KLpp{wdxc_l<lnͦޯUMǛWrR~e>%;?6F5Y12bn/SѕtHW4eB9QfKnqІxgD|kTW|2`^Is*~IbiI '~\I$]rht1>խJ#B$MG߻6.>;3W~jO>uP%YEhƈrL/;al9+Mwj|I?-u)=^j@}!x|cest-ֈ25ol"biY54CSMLBL4gY|g_)1wMKfc>__Ǘ7=mxOlmtCl -|@֪Y\28FA6wgP4>OM,y4H)_ ?6LOwoofJbRQZ's[`kF(Ud4e$PC} Bzzj焆)DC@_c* 0EBP(l*Z:U]gT>T݃oe% f5; 3wѴZ|S U9 @U7j~q؜ԛɠVO=Jtv=!3lB- /nc)TX,Q` e +gVѯ$FE!9˺'bEŕ EA\T_hmB>AEF9oSZX]n]1F: lp6HF_j;[,2Vَݺi H͌F6OGRg\]ZO &E5nOخ,M9fgD [GW7}¸φ#~فI%7iQh 5(o+UlGmbPK^BMUs5q)\Z?_SЗ /| -K+,>0JϮdG㯣s}n*?MCpv_ 5HƎ`ў[wCO!.4}AQ{ XJb@BD=j9jJ*Pȸ7#fd܌q3䌛3nFq3^AEE ^R`tJ e^Hܝwpw>ܝwg먲`Y,>3?jb$j)D%<&Tf,e,E(2EƢXJw4C_ UIf`3NJͰ;kH)EВ)E 0ó:L.; 1K vA~vfTj!-Ԃ] oT٤ҤR-M*aƔo{q`LNZz8YKɰU'ӱtXߗzm>z'2drM0WY97zʺHU瘄@9-`BG M?٘;F~qpճ2{Q =IKe J(^Yu`^#0!W(X"8,(" dlhr *sVA-m=3{9A ; vFNsպܣ I{,l8}sٟ= }E1z3SFLWOd iUU[\&|u'.- e 6Y %,,qrTKpglv+pK'ml͞{]9O4 >EKn+gy#jv?-tT8*ޞ -C@S_(Bm(B01OSśW}ߵ-oz5CR gAzT(7'MSeT=UDbp(joFI*x6^𹤶L!BDZH1WOp׆dWl7$g.n+ڸٿ - >WJٴAaTJxሠ{c8eXw*+oS$KJ$D3GpSmTva* D ^9a^ #ecGdTaK,tEEl J&=E@='{UOkWy: ( N:.-Ii舳\SH&[}7lYi5/ɴ>Á' R-:#pE-yMR+e8߱HM0&sɫ:H^+?pmO5]Ɓ\Yv_ Á\T:%$Ǭ޻!>*.ET߻X#,L m` )Q>yFO0e7<{Rb4B3DvT02N%D 5 + 84z %=L:VHYu2LwZ D佖D 40Dt9-w `=< qUa.7˧t6k QK[t8Hm8 0 Dpagk5Z z|s[i&VךȍMӑ2Yܘdm:z bK@ZߟZUp[:ٜB:GZв}IϏw^C?yf~HYn?~sb#~q薎|!,|-9k>ִC_]5s&M2'M?Gt:9 [OD%G͍\F3,D4-,eWF)  ZZ8JV#w@ӗoۖ~ܙBl`tfH%zMq⸡t΂ 9V%6/0{`9Krv{H#+>%8`gw:r{яjM*$>ő(q((,{f?NG.'3 !qˈOjN0$% %(Jh#wDTP$@&oI_ 4(7qɍTyĄ [#~q)r i/IdeÝlnl~gݲr [w%8VzRU[NGAJl}@!'QI%=Gdkz`(sUͨN1&I󮌉SBNcEѤ$R)\B%նfl-݉R [GB&ՅtppejzW,T_n>2 u̓_5(`04?}J'"B1|`X 2pM[p:\8, ˌ% !] O!.2Q#p&jHĄaU HЦe9kl?%XvkX[ڪmCU% H*yEF:F5з rsxd$]εFuk:dl-񹡄(-)8IiʣF > L61h^mXM'iuOI;i ۃPy?'tWJt`%ʖ.ĥz[:x(Eaxy{߲.uC{ �|UӈL =;ڣ=9SRSK]ܽCr72s5o{g0y1ORPr}}qVM.^㕿ɛCKF/LD#4W]Û1F.6WQ+~FZ:S%ep,!*K Vș] d9;zΞA/x w  xp)6ijOs Iy>,f?2R'Z~ѯRyE ezo~- }۫3'9^WEJ,a/=ޗxNxѸWC0>ǁlUIG!z!Ykztcyǐw5_SN(,`Yͳ}X٘זoڬW[>ۅ}OľibNBQL׃ 6g^{V`$VKQc`pre Lq8j|c90c. 8U|!YQ.x+Օ:Y&)WMFdM@ +&I+w&iI IUZHInD:ik**YFcBbki O퀃ȹ5t"Yy􎓨"[^|W}YdEmUѭhVrtZå/c)uku `3p]Tp >3[`)42w`;<)ʏ9ӃOfqZ$. 2+ĵzݺoBg<}{C}ot< Q[qǗ-f3țݙ\eښi0OƦc?,Nn:9JZ$Bg4D0KmԚ3ęy]x=b,9A]`mM6𝀬 99XaFqV?C_3E9Z;:5Пb?xߝdnN0;<`y֝6Ջ V'O,wVeUZ6N6 6l edޞȖm:.t#:MxI˭-wq[/xux-u-78'=sL |,oqߓ_}u\kl3{&>=;I0@[Ea&n'sTʤbD6e G{QP-Wr:NbUJ%L_ Jx<ne0q |_F TJRu:||v[%_ՙW}gg均t^@nsvy9ٺaTR¤0b_kцAk,' o~)%];&vBkhog'^A3_Os|9@ƅ Z:*1; WxaHQo5Z Lz©`x8ʎJ&RmLD2hx5}g+%yU/V?e6&KUIB$ D ൣ/sn#eHZ,uZ0eS,{ќS`#jHa,Z=QCșhRZD!I!hu褩' B7DaRk(c$rq993V)5hI-oܴFΖ'J|~49Ĺ Xr~(дlro~E^n*y,e*"0+Q!R'D$ yoPL2 Ի9$%^ EL2[S38/#`6LBop"+dq P@q%@ uQPy, tK/܃ +=n:$x-eR"p):InY)\VY{CHxE\a3MOeRlLT`9@DuD$hbg0֫H:AjW:h4'@u`_x6sW=8!B70)!?!Ep| #xةT*$LŤ]_ykT7m?be-u[7P,dglQ$/`b({|nThjUˍ9.(79ޤ?B|Jw~^G5Jxs~:vyA>Qf˿na40lQ!$7^m+޹KkV9INY8I6`r J>>n$780ի36-aV%:?1pbRaM%Q6C}0 nP[%o7է8k VQVf8˳$Υji9kiޕjn|3$ˠy{t4T I FkC&m!k %!mu]D̲|ehjp,smPYr+Iolo;%r/J*ۇSg^vtCWaz^S? ʵMJh4zι w2h!">FKDU|ѧ^'9?GzL ;G~)(SH fHiК*p^$ t =|VΎSk\WurHhPVB_u )/1!Ra1D+R Sm"wǼh k;Ri^8$iBkA'2Lvus[P̰d !:#vU;Ǎ4ᣱ&A omDy M42S(y[IE5tDjf}#[P>]>q6Rݣ}#C{G-+g;tt tdO% 9A ی҃י1{T)[ڇ,P7-̃קt*M C/zð7v~Bu\{yލ\j:w}i꫹}x5}6,d7m36Qx{y_L;wvNvS6IɽwCbm+amo)Et|=hc5"Cw+'Z"Ü˄6_ 'H-uo;SфO}%@3Ҏd5Q%3nv) Ձ cFRB$r2y2JL-'ǞL2}y$y+6ƫ?} $)8Be 3W(Jƿy~]a 5Q+2, k8Np)JL:Dt(^0]vN)-ֽڛ+EA#z){®vkE4ijCS%dAq !5)bpJdjr+.ս)B,7y"8ߵ2_kŮ,lŕcrc7%b2̳)Ku )ix.[ptdaP -HL2L|[\u]v(5F'RȣRFe:%DG#lFm9k*oI+pJF *Z^l-6Vz#SΓXZaq9Ƀwj[GpGm`gٻrW<"9Ӽ_ 4`} ٧`bՒ[2>E,:lSQ`mxȪW2C2|e[m9G@dG+n2XǞ=(];YQ#l|̚-!AStgRQ`ZLdF#)" TKK97ޗ?RRt .~BhhՀ9nwy5W[aމ~t2B.޵ E{^ KTY$h@똙Fd FB-9I&b{MR, []E~wl<qԛdK*ID _LI\c.$QY]/ID%aI"#5 Eii7^p\H0̚QM2FC~4MBZ}su+>W_O=6a[0p8mI14ګ?? [ݟ5@0"&46J;r#1fb  X=s8ifORn#&\OZyL,wc6gkԋ[w9a<.;z{6-3Ks ?uBˠc.Cnhf>zIWcG7gȆF6 j=~}d'_ym~z(I-{|ch~-roE= Rp^ڐ6&)D"19CI4-Bp4hY*<SU2g3b` qDOΧA`HrV&4V!--爌ܽ(,eP UW “ CO4y}cvo|@; bt%y KplI `A #d#b#+Y6/`$EltF:B+<¼dg|R2UՖsXbxkeWv5xԶv`'dD  2K9$$dYkz7YeE;bX&mP$t!f2 I6[c7_>O P'cMC3AL4T`,)0G5]JZsea9KPKh-Q: +E?*QZИ&p'!7oY&,JuwQ%\y:__C6M/@ɋCPA,`irW#@K}L7U'YhҒ|h.Ldpwݧ#hgZg;^[0k盞o:7ܪ,M%r,{V M$ōƀ[Z%Rxߔ gm|h/SNpX=L{kшÕWT -`C. n2`wbHF[flQÏHYMFc8]Xb5s2Y]/08DrgN]vYE@'=z)Z̓5zOt[pR6EX}r 9YWƋ)iX):s>E p8P5yd9.0!f]gVբ^)Z6u^V7Jj>3LAO˝^B& i_ҤQÑujc49٨a~ _]~QKUگ~9A Қ$W"ɫuHu.F>JPKѵ>_谝iG_g~ݗ`}}>Y\!N6<ׂI:zn5kTN$T#Ӎ#566pzNr wU@ m-)q!Jei磢-XhZ3z2#ӢVt{n-pdZgd&H==[Ǖ,c٘1<廚R&VNxD`&2h[O $/_uzÇiA[u gWY4IKZ-sT "1*<)oYm*Z/1B2XeXsqFPD7i55ǖYQB' }dVd`R0E9Cg3g`X,?1VWQMC\v20V DlDFT&.Có@{0zFH/]NCN$ WgGtIt"4!@r.JƔ5 2ծюdM Ͽ B衆柹I3У*&UooW!M>neyu]gu?~}I4B]2v\-F`e4q+[c3ɸ׿hjY/6y4]1XqZ~pv"WY#L?١ Ցz4NiW yp[qo7^3$}ytҙNE W~<0&4 t 1j$ϡ%-_OKӻpsCs58=%ѩ'd6c)I1AIt?$Q5nQs3Zy< "Z>xWb/Aҏpp:ŗ6c}Ɠ\}x쒏bRyOI]krz9y҂+/qQHy|f%:#{3~sv0Ϸ9ȝO/Xrg| =qjnNYZǔi=m'g_ ^m a/LU^W\tqZoy纨B)Yst-Q#m {3IwU.FhZ"JIwb?dYF w`;;T~5xQ>)fjg_lucq,%Tl&5Μ>B~F;>6Z qcʍ5J1@rr:YE,`y(X~$X˃ XvCH'Kw>^B}H0OF8=1bpUqRt'U{+%FEW42%AeTm9GaN_>m/)e.wnY.7o"^=:::V*V̈́eL'?}lchħ4HIjx 26"W:8D]zx0.=g`;6grJ+IgqILXa4*R2N` ZL ssFC%o2!59T0^ \"V[1grK}k_rǴ<[võO{8Lp{s7o[hExLxUIgxu9[z[cHǣfT怒&6΂m@Da`xI9+:cE'`,3RR äFS" ,1sQ9˥ŜJkgSۙ&s]*VGw;hI\MwЦm2,wѪ*!Aٻ6r%W`gij:͕GzsI'߫yA^j7W^L'פZOob]C>=wI!"$$&?:Z)X$b4쓍U X]K^lމ+;cvdj 9wpUZpCxUݺ]X߳2֋Ux[O..{(S83Iffu |r/vȠ1[FW4~2|{lMWc6'ʌal7ܺm7M;~E6r3?[ݼO~λdw~ڟ_Ƌ x\+\r议Ew1/KTkG-SONQ!WS)GT=t#*T:ї#z刜#x+9ƳUA\\_h1$ ITdyJB"6"ukt=#6/po_~ 2[Zye?bpn{lDa=3k[S ~X_OyN?~ 4g4kMo%qE?N^gج=@A%*fpl)EeJfV3t Z tD W";&.&-0yyL2 \fc۝"gim4 F`>"|r'= 9`lά\Owp˷Zy{^&& VC:%^ ^51J%m-fTOi@,IXT{ΕTۚ5rv݈R [=uBz]xP]xǃ~yU) 'wtnfqыɟh4:|BAd2Rp\3.y}) j}0B@uD1B4 (FQ&i%#4!YtI:c,۲n;LƸf\vEkjmӲֽNrE8G.5[rAȭì1X%YȄ $+:B2k0YY!dT'9am/tE#^# e,gLIfkRp@CF\QI`dIsH!"-Anc׾v[yI;i ;P pOǽv $sWƶ_ dz)uyFץ8r鋣#T$\J TeN5&CEN"pH^5ڬYDFg`X6Z#"N\A{Jq|T08%Yf-E,B0Wm{Fv(IWYUb_};{Ӫ^u[ϫ<=1 T¨!z(jW#qahlF0ƣ֕VT2\ \%dHJbo5E\ɁEn s}UYQlMtE$Pa,gzyq0K}if܉t!ERlHem3q \8-[H"ӑY3s+^''d0l^;#kr9nsŃ/i>,jVX^z* ^nz},Csg5&gv 7hш핬)*x\\G$^0',^BÑݢްLԼ|waj`]W6n#IKj۰y\;-m"ӋAoA't{-%-wo t?pw瑹wS.VW4H013X>][r.녈=f.Ev2Lav*C\f4wKTl&eyK/odi*-օ6IE_[yT~UYq8qnW5tJ'%# gNxLIIJ̭u(C|Zdr1_sj#SXΙk/[/=Q3NsǨR &exFCbgKk1`qsKqV~h8^skΤ*ͤ$VV',Μ0}ns 2d] #?ZfSp/Z5WO+fĻl4!Ƅ1I9:Er"22́՘J́,2Fɤ/ =)S9{nyB]s _N.2]XF7+WVLu6wn/kz[W^&-?_rx>mοh0׮=5zѡ;뛶wRmxnlsmA6ߓs~H*Mop(.>z& t lb 8A;}feH99[dg/;}h ]aoV QPAIY8`![g1&/DFg4-!9c}Ӂk"lx-#@rcRlN1/x]" m8> ?pex;}5sPRd2080׌Co/-'t$K>If jlaE4 MQKBGɀKD̺ ҡt:G,[YcW#gQZ3.;ywTkZ[ZG8^1bp yrgo$7lHID"WՇYicK rIVthH$dQY!eX$a5rÙQUcW4b5F4F5N?>`U,'R$emH,ŢgB"Qhc$iD!֮Tŏxfl$^brhw bE8#/,&nWT*wP9_8g R* ?Z6*QVGrI%-3:A 3v/D*! ΐU'<;bIaX6ZF1bȵZ;Qt_)B(XV*l˒,RE`1Cɶ̵jW#g?$ lh.%+`pGǴ)5z^}{31^u[ &>zy7R]P ="2u1% \]J B+Tl_BҪ,Z*{sf[ ܛ+Q֍Cn\׀YK].L"NOp|}]lM fbÛK$PX& F!,&Or'@BQ,1 nCBE2NHЙ%MLGfUZv!-gxZ8X]=Y`L;+zKqZ=wkfMrM'Agx*%]2F"QUqMJbG!7ed[@;U<%o25.{j3ѢkuSTJ!-nF[gYN{\w/OHf/`O0,^B[WްLxyscjaCwNJ-6l \f=Jm Wdz1K뒖:5ˊ*o8]My"AZ+ 5hFPbL oNLbXwq充9ra?'"xbhˬ :@i4јt.Xb5s6Y/;杺Pim.D@'=z)Zrͷm%y6(? o go@.]f֡ҡ1Ag|BTҲRV5skz!q|䝓)3sX`Bdç/-- )-[6u^W7ڲJjЏ}fЙ2N֭ޙ=¶i^pHj1V4ҳa^,HѦʅ(r%7:*Q'q1Dσ3\, Lp XCl0+Nz<č1Z !1g\WɸBJ~iŽG g$~v_n3&C3AK~ݝ%$@g3ϓ=0Eq`VIFKfl1u;9]9x/9xD>w`kpfe/;G;N?oaK$Aq<'R90-j睁Jح.JncC[묂̧IlNObGWa*oeY6?滺!@ZRKD N!pL29ȬD ^9$hL AM~PpvU5 "Yb(4$􄮼֑`zIPG`oJv㌡qqc&zHP2`( x!k&Ds(}M0&Y~~kz6|&6[HF DLDd&)KgHLxtй K:B^Y2#0@"L5DA1h<$ 6F4:=>^&3R~Y@近ja޿7*$N'p#0:ׯN>lu 6sx?4 la >OS9XN;gq8^ b氣^t[`TNg+2寫g%AHdxWW8Y~}Y-(QFrgjsKMO3ܔ^A#8T}c"Nq!̤10͟KJKҏ{ouyMNR NINқPNLgpOsr[4܌#DQ3tugK3T=ytߞqÅlk7u'RnK>vRuɕ^_%ytJsJYWN߻tx~ԎYYe i7胱.%E1[Ddg2Y!*NEWEaikF쯮Z;fb[]BŞU}yĜ N!Fl.6J]? qcxRyk)jZ- ,+˯U[~|0VY :43@G[ <5XȎZy^Xyd2$C="@\-<%md˨ ls@/HyGl{I,_9}ii}c>SͱGGAGI/ab\/Xֈw+T:;bYXucnJamv;F(hBH&}㲈 eFdvbMŘrS7hA2ˆ(O6@r(crABnЊdi  ~-jE*o'2q]F/78<и?~!pDn06MLw?GwQKT~ܸDN->xxi@q TUW  'nm݈۹LJQdjcQ?nyPJaVRn,ӶEFs~#8PO8ow6 Re!&O^ cZ+ʣ0,[!iD:q2?E@f&7+x[q>*wzG(x6PaJ<3(rI_GP hSRiE,+ $,f@k}:|`ݰt}s&rF1qio5Г9JtkgVy & B 9Ό[rO>%#{(g1L^-I$)(K2!t΄HVb JLϑΠS-b4л{NϿ)lI6 IncY.G( R ȅ`+."R rVCB @dZ WDabU!KWZi. +%5uma(f_JfjΜMǓOj wUHܘ?nWhB1.j]9!+녭Y&sVJ_ܝ+hD!hmrgГF9$ EZ5BbmoJ_ ok2tjvz?ܪ.*%кOgK)Q)jfT >A +Vk_)UԐu}rW+ռ|%Z:bV<[QE Fl*d6Rk+[ Jq!rAZ"JRhf:C.Zs,p-",7_M*Jn;5r% ]I$hIŷ۴*_ĄgHLHl}~|[ Hc+`/I`Ŝ?r_Q+\^옿nt'͘XmK#Xx^7>n,;.?G?J*_Fg(^wAq#.Fܒ6uz)$L3*XtL*C)U# tLSǛE7uEE[MSw T_K'L/W)T&bDB)@pl }蜴|ZIv֨KHs6` /E0$s %$rR pp\[Lh(Ӽ*NVɂ>Ktb ;$; )j)34%P=VGrl@ExPglV'QJ`Ӧj=UdP&Q'ΕR;"(P$Z$/XLn.-KOO1)(J\KNd=}0eF{ΒUVS YdDwD7,9.~իH) LAI ]ԉq z2: BI~շOֻ#^Тa? =E^mn&߃ȖRB,y tާ\ҋKWu dh5;V-N{uz>I>/ۧtLߝ]S{9]O ps;4_jyUny{ͨMmX36i}F--v )Dz|y(\-u5s|}/mʼnY$g(CLУ$DhQAθOƈl%<2zcr]&88a|͖NZPsбxx󁞟6!X4QDuYy SR,+2G$TfWmG 7Gq6yzYxc{ד?պkmA!-˪.RJ}>OPuQ^QVO})Ŕr]X9f$}vusbJ.@H c"}D&wp8L\`Gs` Ffy@z:&2,`:Ff*9$XZL 5ܥą`Q z)U"EI(-2@o1|/v8]^JQ2z7XO,D+ /aZ>滼.GR&ZAI$)|* Niz^{e7VI!1EĚ#HB_u S\hhnr`yKi .Q20( ZAF *X4ZBHZ UNz:,A]g^)QfN s|LZdk !֯YؿT7跦smM;!F0tLF#e&di0ȴʢYD"YudBjw!rJ:N9#0@2M9L)p}B 10:AR;2hhDKv*E>KN.> j:oF՛Mm0N?۷6awj~5?̎2v_©P,52˙d71ZͿ q˙̤^WA`_&.zF?/qL_zRrJwu533WWjA1k}{WC o'cnuH;;0^r0G|dLiś 3֘I0%%%x^񻻾&_\٦ ~+` )v*c}f֫jV5t}%Nƥ_. \rJsyRJ AET0Hj-ctWށus QRj֥#Mǝۂ4 e6%i$KA=ZKh)$1g3i>Vlgrjk'rj#tj[o ZĖF7N7GpU& -gR{My45BFAIyէo@P dBdpL+Y{c?16mh]6kidp2B AҰ5leC2v)}-Z!Y |87Yɳ%&Řc&\*lWsHKLѴ:;Ii9WC6Ij QidG"2\"GЊHhFcޭij섍1'#W̽*:#8PځV ,|`kR]*y`feVry;| &NU: חSD#]030 bzLxts(T:L&7nnAmZh7N_͡죯FUv1qA.D0x1.\.ſPD] ./Qh$ r" "ZyUR^\@qe釙gW|/5<ڏ\cW\GWR_\WLDݾ0(7>׸d%|x*4T^"YdD-23.YdD,2{L٦0&e#,fC搣rs5H41 )ڶ d%g=%K>&e4VЙ ρHe ̶li3z*j|e Gw雛9]wL<Zb~f,TVR;ʹJCcr)k!<:(|4NYϹnN=zgukkK"2 U'3.9n*h6+M[#(,]nJvE[C<` m)j7i$C[-# jk;$dYyyZ4hPތHkQuEH+kTQ8z>iPv7BNyMN4-2 /i;B7~u.ce] ݺp 'Kv3܎etF՗ann-TYFQ[Fz+rj2qD~UtomX>ckWb aF+@qPeǢXZ>X5R$ ٻFndW}Ŧ$WlX /y$(ɞoud%Ynǒ(6YU'.; gpA<:%By^ZU z|J&]J$чดɜV}&O6A&-R6/Aہ97a<rn{]'n􎒹)=vY,1{`)|ԳvΰՁ#]}v".mQ[pme֏vA\L4rqN&݌NH#٬i=3 >YzPs s%}&!շdm9%~-Udaoq,_Y(Xg;>]Z~xdkeWU?O&'9P&ڦRpf\&f}0B@uf'%D֯ Yc(G$"6I:D-I;B#0벐.IsLEcF{k9%vMq˸:NڦgԮޞrD8G.)jsAٟ)`d!f%Y!YE-ȀJE 0+dd|= z[ayk ""{HDS%b{)Qt ɹcT!LiĠIhH3@ /b$II.k YE$I'MkL6Fb@! uo䌾sqF!L]ʒ*K3ܛ)zL7/cfR&ʚSZ/|6yfY9*TI)5IHl,Ch?&w̝/L^J62y zh5Ft7ugSv]6@GTPTU&,6_NIkf #q q:{]tPp!HC&!Xeį@%I(gHmhm9GfQ>TA* S%I 1 YǠ@38:]譐hQds{SH٤oQ(ѬEhnXyelf97)`qLoښ{ډ>+$:@Z|FS * eaHJTPxMT&ȠTL!2fE 1cy%09T}ռ oN/`U܂{AJ@^˵-~'Z=u6zډv"?؉|?ºdoUz%M8!E k k kuuk VNxxxohh idap²\oVUY ApBb'RDjHy"5O1(8tDVD7qk?/H㹠x>ٗCJxêϿ d1e- []2KozɐP+Pp3*WpPhyE fWCGk>)bʃQE׸v>`uy9 :b 9F BsЪ8Y&p%FKc~ޖsX0L!I\t7V)[iA|pD_UZA#w7^Qi~긟fZOtәꂍZ+dR3-N+rbdFd3TJIǞ5oX7pHG KTI h4³,0,Ef:Ek䐂с䔃*y B^?}5N7h,?y r;@=C? z>bǵ3/iUC.x@Gu?g3w5c[,d ()L;*[HYDW/O/WOn}qW'AwPO:&1`4Bg0e^DiBV==q1=9zP|g] tIIeNxLIIJ̭%YF,E Y4L6Nm+_pjU\|=Ct=\S%{=ך# ˆC=Seub̉TcIg\wsI^rӉ;cH?dZX4Ŷ[vyϣ O)QxotYI  fl0 "4f.mlJ^1Υ^tCeI5ܧą`ЃmPV,3t4gt Xr,`RӕڭEY86tM8/P|t\~ؖ-)$a{D2Tl`2:, bJI!1!lRI[No4qr(/* |L,R..Q4V-^f뤊 qDI*`>l"bqOǖpҴh<dƐ,+31oʁ#Jvf<5'" ĘW_|z4tT摵<4@$31}L\gLxtV$FV_F: Juh*D⎔th=L@1y@-,4:ǝBݠLvgDiJׯS?%{!q1o+_hZ</iG0wnl6kw?/]OrV诌~F?ġ{=Qq3n{V_ [FԋÊ[hy,vGF^SLvS~$ْ]^.Cme|COh1wcl՛ݍ9Sot2|(0gPԹhʯ[ .$_cIFLiZR|Z2-=<+~WW?ח٥$)v?$;@KIiJqJ PRZ^V aVZ%?o5Ζp 3P+ K{/yG*I)3.K;vZ#wZu [ oVgQ\=^ضOh2]?>esKc)9_v~dϾE\>NveRwwwّwn?<%޵m;͔8_&[\?aBf^3t˞/Λƥq[Tr0!唌9ZNB DSp0xMI?vڡc]t}ix(y-Cri-dp)L,iƺs-%M)Y2,5vKlPImIImdXcs"hO;tI{Dib&,֛>/(sFwm$_!) :^]`1`ILv?$X,5HlˋyMR&)yI5=rw^}{u*O@I kG"-LԐe_كC @)D6i<\Sk]*NmRl!:]e c |4DcYO["#uF=٢ڊ@:,fQ}^Z˃.l|yeX8f+=T0ੴ1^G'EK&2mL@} zV H ځO.PPP`>G>Z N}⥻#_؆Z̓'mL}؇,!T =kZ7LLOt%l ?417\9qC 6Nr#wW=]K߮_FF㯣V蒝I.į9fu '1Y~єǯÞK?,v kjZ:FXRRIʘޟ5{*2R PfC! ^pG\OXTuƬZ =2 &Np q,JI. M# LO-"\l=ЛJZ :mh?}m@˚tc$r M"ʤMJ\O#X&qS Y;ü3B4.sˏ PM|[#7\XBxMI^"ۋ1 \OEGF:FL PK~t>: q:_qC:;K &I*,K1!O猘(hmOXkȱڑw#w`Q\>~[;,ɃbɞM2ℯPHPVCFrӛݻfm#$ J&F#6@"_LtNFub9c{)LZ(uSTZh`xmh3L /T9|e`B^*|֜Yhٛ->| /SO Ox1=8v?b?̍^xX^#f~\ژ/ NFV ~*om[|P'jQ޵p7Ʒ%d0+J j qgÕѡS9AOW5<Ȕy@պuֺgn~BZ7? ,{M1VVqG~4!3Q: 1 )ھpTo$~% f|L..GHDPu.S۳9[Vx5NIޡ*;7\b~T?6CkLRn٪cwpusX:P;o3rpF + ,:/)H|NZϘ޳omΪdS֖8IVg>1M@cZs.B}a99Qh&wf];[zŞxuk^Q a1Vl a\n+TrQ a H=#q3F\r**}PidWP\{Bdg# J*ړw3"Yz mi 8õnc< r"JƵ8KYC&2`P*ZG&3,DHwu~ijW.]AwPyxL5 PTWcA=x)C2FJvBˈ%| dp $ۆuԓ 3, |kECTtoCS}Is=.*6DF1%_҃nO 'em$aA]"/]ʾo9 ⺽w4(sT`v>mDO~ t%]/9>fxc@ @ ~Y_.o~zd(dl)脃2%)hW;w Vy) If#gO19XHo4RLyԄ EŤinjGhv0v2ca'cg?K~8!Z;"<g[QhC9q91jVZdrĿx@{.sF:"*< 9zcڄʘ(zR0H<Mt)2 %colQb-J9,mg@p8:޼pm-")[iή6k"`"j@ˆ 8 EsSA"6Q:6"aVPc3RA%vol~Ca.pe,-N3@7~E`Td!r0K]F2@ !ԢHeML4( d@( ,# T'nE9aϣ`<8Dq펏\y1- EI(T$5;@+w: oku$A0f"jJ iL&d@Z A7rh1qW/DTYReI}{`#yf1q"jܑBJ.{fAJiRzdqqSA!@")O$2*ɺv_qq`O" UEQ5`AhR1zgɐUbJ ׺/%BHu),Y[ Yi`YcL{V[ۃFf(<Чƞ_@5ISR}LNmEGGFZ{N5fQԘEr5RcRzYf<յ+1ˬҶ0 \PEf0kZ+)W[JO&?{Fn kvXqSNʛs!pUƶJl&c[ݢuU?`M6<"5zb JK)sob`SB)XƠ̈BÈ &#ez:,ad"Z8:Ly>hBk9KVZ,O3uEﯖzcD[2hp"`:*^M`߂csNzAHF0f}cٵicK&x.Bsv[p(`hSvEM{v}S~^jˆ \zZQr3"&+ndth<09Vwt8dC%;3ޔ_:#mSt"6K}Bb0ߏf`FnB@_$gX TnSyTPLMJy2[LdNXIUc00[ymU.d^0mDeﮆ@Ӻu1>W#א*)1€ D{ioڻfK{$*+%Vd| kwHAXX@zr # :arDƠ5 <c.4&aT֓,i`N`$m,L }3;ΆS-tuKva̵ ]lT@c};M^SwκեLx}R4tS0Q꛻}գ hrZn`!j0XU cB|hgvu넇yqϚǽJ9 8Jxdq钉\IGlQl^A.x]=79g%_^xE'ߖ1((W٤)E}.idJk5=Li]7FDDt},k)?V ~$9UdD00YXό gmE6{nW124Bj]J\5 76 Op i5->js܆ƚqL[u%ci@@iI?TS+8oOI%<^) ,hU  @FI,9!ae3.KRTGiCb9)Dڻ(69$9]f'Ku}OUfP1zZBh(5s&M@QmjEvc"ŭϝȰɺࠐkYul`Z}c +c5I2qxM"' m5 ;hk()KTf rXP]^^1 "M.r7dW?_{C+Ż/*" c^JBrZ@o)iAޕD D] i4w@=>ȓ!ͭw.B{OSr:.\YJٞuX2NaUuRmRۢ-kgR_dQyרf)NJ㻚Wm㥷g+j=ƭuCgnQ ⰗE)Mh<|<A'"clI]1xE-T6mL;[Ω%2IؔX&K66{c-DB!!g?+;;kU\^\X&tg촳w[:]n>[' Kz_u݆Jʫ1N&\YgLdqW(Ȥ.=L)gAGYٳc{<:,i"k5I+T0!֥BaFd *-C8)qy`裱&FAom%QűCG'BG2gC 6ܜMOi.s)U1?Zv1?o0IR m~ss\zG#9P{\=3Uo#G9>gA pjx}9w=[q&C!h܏x.'Wvdr.L[*@*̜ W"ʨ0'b,zs>WȨ}m̛{؁.jK&%*, Hekb,` ZͲV*@%XpA) GH#!d ݢQfa4dv\~o-N)ڐZO&n34('8@k7se t`@`Z6u`%lK 3et?1h\3#h:@!П0ńpkZ AXI(>C3/xQj:t~ ~=UvoZnY7|^z/-FBJ&Qp&6s`-i4G嬖Ek5n;' [1 I8p."/]wҁ4Q׸.F{<9X7g<>6YhϬ Rn1"$ꞯml;j%+ZDV-=ɑ OiMddBR\ƔP;1(=3"0"yÆ:v>CX?YF:Ly>hBk9KVZ,O3uEZ1Zb ʼn-z 48i0H| EU&0Bo1A9D'= $p#tYBEؒI?bŜ4 Ԩ@QӞ]on ,C#KM2BO7M_ZFGscmrN^[b:G4g ǐqdgƛ2Kg/Cۤ-&G&maѝPMT7OmjAzHEY-+ G\,sNG=՞Gu7) +.Yg**2q[X&jbF* ok4U;͂98R e??2N*$<3b"dBրц3"[`ʡ7L1l M pb#8qԲsM5k vdF]IE$Kp1,K27cu;pTi$b~2@scZ[h4#xa3G Zyi"[$iI|_?0:.c*:R݃U@oLK6w `X\:.T;~K9>]4,goh7Q>)<[;DU{6huu&Oi+ZZ)l+yE $].ID$K%DtI":gn$QϨrtW`(DQKPW蘯b{d$ ўH#AY('Vs @KW 8TXDIHtD ٸwl^ԕ]ZjCsO_;s"m 7}^<!_fUdDF50Y6=Yobq$r5اD 1U[K^k"ݚLy2G1ӏ>WjQZԓ/ƷaMQP`k`e,Ȧ4yry.8ŵT8qp0[w/1ؠC5ڂKVMI1Cq:440l!섽rB;C}{^ !rvq9R_~>xyz^~{ҥi{VcQ/Z_сY2No?iL q _||(Άd5|/Yr옓[[ +gᤴ1se,/mX*?P6*(p`P}Q(Gw/|,Euzx c;[C´T&ij3Ɩߖ͖> \$6- ZXRwSK"ےܨdIBu̹.v97^Pait40яV_MJ?FG'R`tϫuWoФz/KJD.SZC7jځ]e Ul7ⳬOU( V*v]j?!iG''ojWQڹw}zg F7O \%-jwe7\e0?_rFyeyVI<ֻN. [|\-jyìiܕ4n{{d}Uc]F+j[WV[-,q [6]IjE?Q1Ԇ>M\2ThWUMڼQdӠƔ{M9-)إ'UM =_,kg{՗k6S_.4g_:gm;]m'.=(ہL"0U#8%m]X\$DH۫|vW])q $&/*׼`U!W%Qujَ[hT4PIZB2FTU_s-S-Q66`[72P(.mן^t9yԖQ2Í>wxH.!?A^>|rRJ7;~u|x oz/wIo={ެ8>CG1q{1NΣxj¸Yb^0+0 b#Ɔ˚~x,?9?|o_}٩g7g?Zm q;]\W_hRڹzd5K&G^Aƛޣc(N^yG]=orVWsGJ~ +ℵs'v@7;X}bn؁ v`.&YC&r9& Bmo\ZGغvҖq`g<8sLh,W|՝-WQ Ch\Z7Cq/%,\p\0+Vl+ qb\-[#r4[U?433K~x"Dس [}ϏN҆G' ){::xtB0!`ey~RN7'0 D$x'(قM/S o"G kvTi:Q8Cw|6vjxrY6Lem㼼ԵG{WcF/pe*}b~ReOijY焯(00k̊Jq)X&_-Slh?[޼?dɘ޼ý-Gw'os >q\VBv~ʖ#C̿ 0t-0j k,!cl|{@tdAPK1!k0fn?+mh"}ΘQja{qLOÚ\uWCU0GFRP4}i%X(;Ӝp]fPKv1g'5/L%Y35dSt[k๥xDk "J`fYkl_h`E-(xM]Zy;#(YgJD D*)aӲɟSXPvo(*+U` ăւ&`:^v4H(g5$J6ʃm %nd,C^N[N G\BY@\Z7(Eraw E : cRVK" aH L tƔXʹ @fs@X + I@S1DC#)#M VS4ȒVߑ@)*it*TM ]==+(EhߪPZ`W5l]SRH V&!"p@SX,]Ȱ hh]vIcD ~lh%SFS xf 5fpҭ݌b]f1딑<0`!J(k424zRe.W .vn&cZ)(}tlcV$F@:)T̀2lxʓia9|Ѭ‚BWEz@i$* iy@T`FRcGA^RA$D _u5}cx$^ 1*^JX8+R~d}g\=U%` 6DL`$`º `|ݮ8cC'EMᘪԋ +pmHizХf~#Ŭ<$D \+:@1;w@ @0 h aaYW[ɱ0ZЮ^1;6@ 2l %aN K̳bF.H ծ5Wbpԙ@F2s>N l`: V1cHPk &Q`9{`. Rb@+U2 -Ih-ބ~ L1C:b8u`b"X#yg̤ m; Rk-fqq٪Na-3!Jw-d@``_}dۑ%A03i 4PЃ3雓q80g{-˴0Iaj @#qgKf/#̆`T Uӿ]TFԬ"nR@΋E%a[7<(G wmgst:$:YYN2 2P#P`[AH $7z&G@=z+ey pU>v"" 0e| f7Y.ă tAkO bp0Z@$y9TЪSew@x'Q,(B)yWIi~8{ѡo<]c-{8*hS']E-lxk`i gmnv!>yLCf֗6jE>j0{6E\AϥH+W)WJW3f}Ibo?~Sie`u0rEa҆KVF,TT},Ba?Y}&$$j&3*'N8- K&1r 3*FxE W5qvc&]9[s%<7p>h3IZvwҌ {r `E4kҤStD'|Q& eۂTyV|A 8x y5]pOӞftYW/فDcP xŮ16H:x"r$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ IH:@/CΉt$C:q@ֆ 1ϒt բ:5||1qqƒ53i<}Pq`"!AӾowx`{W&rs;XkY= ZCy-c}* |'ctԨ俱ZoC|;['{B?]+ӻ`#y~U_A(?nS-{Ay J]-YY|u$kd8/^A [w"%fL"JQlE PEU\E\jm1Li\U7.,,:&=eqݗӓ(h.p\mgPK 1S܈]4ʎ%LǟG~)Eͷ(trF8&hMB;?}LnO.StH[iA0IT~HaL2,˶@"JꜪjk$Wm{)Ob iƑj gtTiaKQjbI^O_삦Z>^'ӯb[ J(m[]69] 7k"2c NqkYk(񟋌p !z5IX˞wЦŮ&n퇃\ծZm]j+nxIJNO6P?;N(61Ks)7XJCM)QP3UY*7 3dDb@p t cӘlAX Xjq*QEDUcSRbcR AkyBtLRIh)PÀ Q^"Vg8}ćBԤ2mIYo= X&mnkZk2g9nع^tVd%{%tg-,{)f?4[%)Blf#4Fh6Blf#4Fh6Blf#4Fh6Blf#4Fh6Blf#4Fh6Blf#4 )]3Cq0XN)5 B|}cBz!t>U\ rʰVq=*#$z$QtϺAG}*0u9?^KVm2DeܔԖH&{aUkk,'Uq&ܲbhxW A]-r^O/ZصWiЛuF=[6 wF*]]?׃wA熐xg{LO紹KZ6gnhȰ{i# "FD0"`D#"FD0"`D#"FD0"`D#"FD0"`D#"FD0"`D#"FD0"e%jMuWs.jv5EJ#\kब^s>\lA!\0 /i"VfKJOy Scy{,aoÔyNU)+[pfq$&ɬg h0@,'9QfYUhfdj -G`ty4wM\@j'dk[]|Sf`S9k,ivGYxi&W:lo!6w5#WMVߙ  Ĥ/24puW_nfWoo^`5M6lں^vo/ >#s\YI<s*~H),\= RA m\$ ic0ARZ{b?\q4WGw! K3b<G~noJ.V?ߧ4ч8|r#o97uVC{7~iN p_F >-]~  {›[V2kdLS|ߛFzB+sJU*yӏ/aT͂mשFgb&(==lRB5Umrd_t}恇R[ITChh#5U4LUMm}(>"qVM;@:kDZ-UMʒ4"';ൣ Jat}%;}Oz7%e׻u»& b{Mh .dzY3Ӏ+`[և^I" ٕ7L*F2Ok ֌z`_&~EGn`=heTNq!ăiᒔ*uR^쵕ZDWхXڻ8m'!ܡqD=No`ңXIJ™,}iuCۡ1&d9@DMDdeQ"XUZ"Tw & u`8i 8p̃ODpS>*fɘIZeyMI窬$;>7Z2uٟoպ$/qHvٿ۳M^if`[/P^`]rxg{VܶÔMضmk齭29d_0nk>6bQdǢrS.C9BpOi2z'N4DR"x',Ƣ_zQ?S?Niaj]PMd6N̦*9_)ZR.`?+!ΪMft*9vtrud>tWaS E\ߢ4ގTܕ7_:_=>.`ַv]E#o0)Nsi|1L=R4gL}a)A~Mo9atK)Lli pֺT6)fXe{Q˺u5\Vg_{ LT.wl۲֞{ nĤ!ǏI7y^Bfs8E5}7WnL׋8~|;{jWם#zO =>~\?4޲%Kt"X0cx#w5$OAttB.ꍾKd]|p[lѲKp*X wHK&RmLHdUI]jvZr>%|gJˑ_avx^#|v'&$RTjwq)KIzH$&/ |2/C{ƚ~|f% z AJվL0hx5q+ƼZ^DcAܟ 4ZkjtΦ|#[C"`f(Q TRE sgM::ostuF٥nz}[K,i95A4e+Bw|YMyRr<\ ῰G9L?͛9h]~l~~ξLwMT ^|R"RMm)+xE3ENP@m,}o2|JG 4@k C?b(V2s¸N]D,J/ #kBQ GSR@PkS CF+Q\hjV4H\Ir%%*]kg0IE0RL %CjmoFa38īM~ӛ5^=> c-?snh$bfgLA،d$m٠̩S`56ywH4r5}̇J0x.t~BㅞNa.J/u zZ'wgɜ5Ptt ,UT:Dݨ ΃zM!ew911NBR*x1zRKaLTUCBv' XwYモ>s l]_qnX۶편G2P0}^ʥogBE𶖐3jβSBL(U_̝2E!Wg5.ڱu1}-'k/6a!mbB"ϊN"39Ҫ0U '[@M_x%V]YF:$])H*(M1 g k 6f귐 G]kKyc=wul,} 8Ocٟ=gm) ؉fn^Ve z6l20[2ucZLZ(jz1NT,[Bz/7Ji>~(>WtRܦ]y}YQwEe ;w*9m!\wj'0lC0  .)Fϔeb@t §"U!REe7Xw$ViEzT Y:sE9GVV!uW߆ˇ˫%/p%UC;][#Q_{1faސ||A9h4B>WFdJ!FVI֐>%  N_"k%.~ ::RIw;Q۳JWCZCJ٧S=X>~ 'a퉞X=RhO4 =dO-݇w~R*V zk)KB]Ҕp&&X^5`N#>54nڜUuyrv]){]Wx,@թ_U @0'Lk2շ_OE6.W0fek% VvyBJDnu2h7aWE٘'6TsC!GnLƫ5]t-pL׏e%nͣkDF(e%Ӳ$*2r$J$.ɨnw%P Ix J$,R$&XG˥(3 :-ogN t-9mqIm- )쐜sJ> "J"p`$10ׄhJ7啟2rp8Yqb1$bRht,=. Б46Eɝ4J^XIYm]LҕjсyЮK"fbv:K rX5d䫡"?y'n1hGзgyz}eoI1kZeOowu~Su,SEuNJVqBwn++yَVj){ o7%҄PNibBP(-CrOZIbhc%{MQl0^ >Q: Tҁ4:m!g;ᢓ$FJ!F/ H@0C9GU Ϧ8m&^9ԋ>]UӬjvT־NL7 ɀ mhnnBFWԻ>CFk+ wmһwOػJؔhq8Eg3"\40:Ya5ܩHň]BtI00Oeu:R||U{yT)m0MzOe}{r^hg,ݱ?,֜<W?_pĊ j[*A⼅ *gpEd( ϔx7!+JX~o4ŧCYI"蜒&j 3sEU_p|uN|(PYV4x(g# )#>g_ejڹ*~sBdHlʈ^{>>&frAF?:"_4tm d 4BTbq J`ѓSo2RT4$j-sD^"x,VeIeNDvNyZ}xpq1rq#Yb{(b,уef}؅hp Qlm%6:,ְD@bC2YĄyP>+oJE$Kfq6 r(RxԶ`|E̾T0bBJ,J n) \-4-II6z*ؐRP*֢f˲HXL X BT4JuufRj^P$b1hG8JĽ'H!1k?%E%SH1#!l."6+̵"$bY++,kUDbTd4f3 (%S!ߺ`+rK~ħz!Zre(KgW;S4LlxW»/bj}^k*G{{-uqS ]NnWC]D0&vHQ%@c\rl9&.ńRdWChdHL;ZQ1D1b:hǶ}% ЇLhm%MQ($ιT F- "0d mI7=9iF?7z8&=c}u}| ֙sv3H\?#_6 dLNX:w`ElI82/?ڿۚ35 :"(&JF*@5"0$'YY<&C9)5q~4$|HȊ?.rj훝vξZN<lV1j/ӶPrZ @:U|1X/ޫhJe$1`xl(^`0RX] YkTb5`$X_Ƈl:h'gmc(UMYIuɕ/=Y9RL %CjmkFa0$p@?awHo#ʈtqnjYfz7*pOJKhRjK( r`L sJG Ăgvu$z`XgɁD Ða"9kRd=p=z4oxiÈ^XrzĠ!hBRI \S,>W"Q'(E"3̺ -vp`Q?Zkv0/╁Oo,q1NV$ E!X>ZRR.B j KX;]PFb#R!JqZ/G?g(gc|?HtN:Yè`9$-sV8;66Gd'1t.g1Y6"|;j>;mD_#|,\*-* m|qb1)Ca ECg2fkx&bK)B&A8 N̯iߞ_q'<.S7g#+:ze<Y+'j"cY%b5FDY7^`Zؒ0mqemH_n+|b3sENɱ4y-_Q[R,Y[V+ n*UzW,5E&t}s) U9f` TUhz3 % lGFDRЧ&%6ߒ9[ h\ɸEO1[|#)M,5:{+)zuEsŇ!z>s?) dN}_z5R֫,wY"w~m{y# ҖPH^TJ'Ɋ*^hB(1Z+sGcե]عL6Wa F 9` p2 .SqWh v^;_];0pw2գHك_." ֖әj"`䋺QMpbƞ :{j[8yO)[`XѺ uE߾ ?FM֦z,_gc춁=}F_/|=%L?~>PٵzVVWu's1Jt>άB\ȫJ-_s6ۍ[E$lʶPP#K_!!E'*ZSܑk#;Qv<\?a[n<=be8C']3hm6҂gLcSo569t =;xvv }(ͣ-3\CcK϶c!s2IFr^,dbd#櫘3u,m2V$ҭnRE^R+^1FEOy8j#i&֫`ӼS>ǟ]!tcn84ܐUa=5<{ XѧCEAms^M86e`"ʡi؂(F`8a\9@AH,%J҂NO^Y]X/Ru Ę))BD̦C+ \Nv>ěQ:Α-/, .ysD1VgmW@t]&PP'tg2'L2P Q%/5#5OJ;;޾ڝL6,ỹ A=gE]to{H x*/4>ţ;q,aI}sNyi%A6Ybi0 D}SWĄ*w"Nltǭ.!ڻX1NB ^f2t), j{̺BvX7gfo!3 Cϳ_Uq}+-v;w$QOѺ7e/Ǔ/gBލ aNACct /~(1Vs/&SS"3;Zf^;vlʽ`iߠ_`Ґ̂$k,0;2ZZ )Qccm6A_$K ($uIMQ#a@M1 gzW'[gSܥ^b{^ֳZP}v!j?8㹍u7W>P1НUf&. 1z$y*&߇tWX_X߄yB[1"oNDhT̾hRlMTI碨x{ 1Vu[_]cUX 1V%r񙒫y\ZSXH-( fV؃U828m98+wqR[MM;Z$Q\T}<_p.LM>V0FTIRZ) G)CQ\c^Jt1x C \%6X\br*a#$B8SrwތFTݻBĦ0S]r^kG,֜P|nN'[ T+l 6} ̖7i֚~Q lۜscb8q _?і͆Ù . 뎙h/X6,yfM{YXwFJ8-KBb ft$J$ .8HDڕCQ]l_&JI> փ=rfFe¡ъl[m E[e`;2mXK7c2JaмI1nf7ַ_Zwus=Ի؈نZr$DI#\ĔT)!H*h2,h c`z:J{bC$e!/[SJw(I"1Qj_1)j0:oTHSHJaa| CmgMXIYm]Lҕ9蔭w#31Q;[^e Ed&-()b4/Xh |Q(ZM7d֩M$&ZJR)fȈ,!Ȟ:Dc79ҮJ~콸UB7z/6I[{{/㽰BY%P1*eD S:xp@PHUߓ9lxb'\tL%DJ!F/TEy[$C C S`1ظFSkhw;d7_7ɆMdzjW-cSsh8 V0m,Ն|/I~?p:gSoݧ]Osƥ8[:3[A 1U>ub0_*UۘGO_G7ۧv˸}7ȆG6sl}mw|%K#CM{z3jl2%̈pl+1V'AgJ /WgoEPt?αl{7iRwn8lGįߜJ]v헎jWdpH3{7 l؛(i2f!gxunyE{lfrּ=%rx9yO'+&ۿfa^X*D\9-PhPgHd\yyj Muq?kYT)cd )8DMZRNyDʹ%_fAF¼/Yd%H@tn 1GhCP϶*gxvs凓AȆ"1ٔ>&&rAF:_Qiӵ ii3\J61QTk[*>*V~ɕf2@Rc(̤挭CV)(5ؓ*l*ƒXorGO8K޿M}|̱%WS`9zeY\>[BmBʐ"y!XCXDI˱AAކR#lC2Y / }FoJE$Zح0ǎ!-)]ڭaǾ\۶̵Mϵ{ vkV.b%J "RDVpK!łEG%P`PƆ2KYN-" dbQ5BI$V m`m T_+5 G=G9V' 1&CmRZ@akypuFk ׮H>D&v{'g }1VY+Td0fB $Svo١_Wlaɞ|ezNI.@-fvcbEyG m? <\4RBk+/O;ᔵG#pt6<♛! _eQÏ3,9I%ߩ>d&Ft&K.$,l11Ρ©lR f<ungZȑ".nzB4fn 7w_voȢGJvf2߯%zXnY D)6*֯X^|yu} QNEQQHQa}$Z(x\o2+ $@%#L4lF#;VFDSB$e@z%ָSuG~ʬ=ThCqоA=J)j qdKRB2]y/\gg*Ijbt,R̕=B(j Aԉښ^%7H4զ,%R#-NWic hx;yLF;&fCoD{ꏬnΉtA79=xHE :lθM>y=V߿ɌLŖp?Xhţi]epEhCc-2ܖmQo۫'H )+A=F41^ zD3y@>tֽ`dWwa3g'|Nq|R-W;q{l20ot/:WJ$<.=t]pj٭38'̃an~g~ViOUzw'CJ@D@:cuD3`C~8K,ƿ8nr9xD6JW(P:e-*$;LD'4Ăn0st$yq :=\:iw)=9qlWsV=~F3KT(=Im0{h 5H[ ʲ# }}YiRDf8rs S7U((Q H#P{'" YZ.&,&0^d$J i@GC62DY,]쭵Ҵ3B㝒VH 4y8`G2QTx!n2˰Rdd2^J$Dz IF>1j*^"oZdRBR.Ԓ|pC2 '\m p n&Fg.1KF QKT:(Y )"J"Z;n"zf90NJ@W%3MoR5j.QJM2N j@f6FGQ#mHU+XX\oƜɾ59ɞ1o~wݙ| 60`y7̝0@}\V5_ْ9NJbD L4ջ(mo{ ucqn+҇FyZ|_||fILgJ+H^,6*<:֬2C"L>IpICz={KƽQ[exǛ75U0D+qjrIQ'"0fX%ᇯ:_XoTn4ɇ6'WΙBWN'fz/umRNJA,Nםcqf%hVw<̬VЙ;%$%+CBTRki:RalE,tlQ@v$p0}KZ\,* t}轢f3Rtُ뤲,⎕~tBvא#ZJU># Ospv 3!xrqy55}ORWowg&BdLE:-Yf~ ';ټauqo}wI5J Ijn{?ys?T~E 9Z_3jUqsNWdݿ.e8;],PK\r~YZ=Ļޘ;Q)Tǝ&tU!dZ#&%| -x:->+KV믋ǬSī*?$@JIZXJaJR>gr06_5"*Aoyz fjFc:ūy?9Ət=~嗶.]TlZ\UnYih\i*#\.uUII]s/=e*)Ƀc;ߞbЏYϭm]߭`ٳkkgΰU煶[WOX\sm z5mk[˗|No1f-Ne Qjr,%8B|)4Ld&9)~lAU':X:ѤMQ3eߺdQ#]J6 (l wSK}-/bIg_,FV:+5Ǟ}c/]Ыz73pb&6N滙FՌ? ^q+@`/~y5M޼zW'?~zҀ5ɇXXNX+xp-<4Z#O< rxcVc)RugMpm8K77'zu1z/L}BrL,T8tti{v}6ifq{5ǩlK]~t:r8Z<`Pmx4 ~jӘfjXpMSPoY. _~qdɋ=dG^o^h) 1 .Һ4!zh-LcV 9O#iڷ[%Q>ss٨E"v"E1(EK[tZD(Ar (mWkm8UM)GJ;_|^,աO I zxU3>Z NJ}_FSƦm<KB7EP&u]eTLFTʊH^i %@:Hdc/iݳ9[֪͜Zv~}Oynz__oqw;BףrsQ{:P/ vB;nxEC{^o`|Mo Y]k" 916yt9LT7?}F@aUڵhJ|u=d?d>/wd0U㵱܏uZxj<"ˬ99_62lb!grݧ=b#@mt!J`)#DM|1R(a V9H$n=M e(!Q )ߍe*DoB9Sqm g?xz[(7d]jN @(OuٞuVުT> E&fr4*bU mԪFn]ɹ o"bC'o\.U}):%BXXH-[~ɸ[K9,l2J+ eIe]C~xHZ'&i4 !sm0( ]Y;YeBfg%ѮE4TT,լeB>`K"m55Gy\vEjFJmײԶ5]o@ Se9 D'5RbE$PZXz#5#d f˚L"Y2*ň2򅈹,jr1,[~y8x ƮH+;)>gkHK "tMg X2' K n,p4cc2d(IȎb3+5YTmک= +ntxv!x=QkWS{T] \;kG嫨CqV$-.'z#K kJl |p"eQ3 VAe%䨥H)GBe%d"tF&N^[7-ֆsAⷫ"2K{Xn$]W%)-KW(כ\vw?f+%XE0*f.oqGPy7.ջq+)omK_oQD[o&ywG΋ c.k vOӥv%mN,"T ;Lz2rw5{*ro9]K&x_;.=4bvv?.AZ)!BɧiIO?2QQ -u#H.hq2B34cG9/'q6^Dtb.ZARii] y'9'&bꐌw[ƃ s5㵐I@.-L%ʬbN0p5@s؎;ΎB͗gs谠Ӄ?.Pݸk/"j=O v!I<1f ٔOHd\TIYe$%26L{)\LL\4f[0 yԚ3ę.x=b,9A]Ogݹ9<7Ù9sD@}K93ݶd(ݖXf6U2m6mݕc0|o_ٴM!klIRJ;d)j☏^S}tMTMU=s5%I~riם깏x)hqZ^{ hW*_yP2xLy&E n(*z]fI).J 9Z{U0s޽ Glفڒ]J^rĩPKɢJ$.jˑJJ#ر8` >pJUE4Y$5 IJIZ*CR@JYmBٮ wPPB+jÏ2R?pZfW~{#$Bࣇ'Z BëZ#vy \k*5zfK&y uLx6_`}3SSI؁@Oi(X4jJzm1=g(4\ف-A+{cxU9sTc8۽]H׳yQUv3?1 s%Hb0II`ԈdA{ ~VxJri4Y-S/"2-^0Z OO0sܹ((Cú\t\24QרtFsGv?' H=-nx5KNɇp|i # `=zE0 (R@L)HaBLj d!a%zz$b}pX0LhHb wX"#`JP^+O Xiy/y: BfZK:sjÈ:$8.N,9<ˈIgA WaF-woʣ$s}g\YDř0}ŬТgw] ^kn 6G';K]M8|3QR話(ŭNQ %9@<A`Jbۘm|/'N98LM­vrqYebRv ebH܅f\v$ңf\Ҿl4癫m79\O mFM#w)u:x2AM"a6O̾ӚLL][7EsѼz޴-\#]_/_PGƉI`!B GDh 7:5K7?蔒.YDL88=sM5;2up@.~LY%,("(V(\$)$j j"\C;jQc}³n^)n(Zc7xX)\@6-7;/OUmR۞wPtXe9Gq .:Q$iI4Z~xﮮ?Y$I) W~GIjەQrJq3%\tU$8["jZU bbҚ|H4^]}{E>l;֮yɻ_Ǔ7XԺkv;RQVf޸˫$.jYi9ow\iڕ{,5tK'>U[k./w-;ȦezJb1 Y;ݰ|-gZ׷=:ʢM;C[m̵!B_WW٢vS7.wMضrL>Kmݏ3Y/q[U.Ͽ׍Kp]TP=$mR FE+s}4A U6ZB$"x'l>Qg0f9v4g%$$q*p^" T 5|Nʩ38ȩ KӥSy9nbknbpU&v71J 7&}FFƚwū0ļp;GE7~yU{&o^MI8YG/4|}PGOӏ?"*xXw?>Ƚ =9vmE31[ jg&ҝ^]^~h\ [WIBCNmG=ɋia~{kȱ eSUuj._6W~\ Fj~.=hZ K|A64*CM}_dbmըEPآד-Il×%ʫu/Ji|PJy5Kr$J) `\镐HМlG]?=Eu * IFP1QE!2L_X gaBtFu-M@ ;y@ :-5cQh*X;W 3 l־-Q9'7|\,NؑHSwHwd\!us앹)sLɏ׳L> q^PÑ$Z:*1;JW6C!3_(udC>8FeD$0 A8fNr6& ^>nsژRhց7Y->PPPM>'>j?g3)λ)~-:F1$aZ)"#5DG,E(C%F%wB]BQUFU=2  "SM,jB6 TtNNDzchF) S: !J)vI:dȠOFu ug_C4{K@M*%,Q V$ʁDҜ•2zB"=YP Q'`YոI岻1s*KƨN)U\2N#LH)IDq\H`V ^VZY:vF LJt*Lo;aN*$Wb$)eJ3W*Jƿ؟N[uآD\^IC.pcN q/L:Ĩ/Ld$('[xZ损'^绊Yz)M>ji,!:Ē5^H=.wNV:$QaWjZ)CTQg)po<bģ6k{Zy@m܎+VN&;Wd}]]#nXIyE9/*>Q1vy{B(VId޻nl[zm~}?ZlG|Oqe|C -^<(oA{KrJ]_߯3tlf^\r: %[xM* fږ;|Neŝ/uzqFG_6eۥVcëk %YJ]C1I{dIiIyI>"qsb1`X TԍYzx.~z+VbZo L:Ysl-JD9\ӫˊG@~廁qFl$P\MaNDYN`5~jTyÂMg;zO?ye\I#O@3R@L{_zf^m,.;喜uKe-@D)UWZP%"֛`d!低XZd~X>>Z|QGta&?O#w|y=?O[F>1,0| *s !fJԲdkj潑E龼G!): ,g-y>%u6o/kݥJ fj]1IiLF!,|TY'}#+OIēC`M;rtRucwc~AO0v$(0VZ5RCɃZ!U5cDd`˵T(9]~\דߑ@;@~K5S.5~AMQ>K2*C07rSLb|㰻ޞrnZnVcUY|Lk3]>6V46H]HIِQ%#HHĪYEP[x%O1!T"+~k'D֘,SMȹ]Y/@-FuS\ٳWB.xb AtA|"ҴT; [ƒmeU|= 92 ʱw߱s9`/+v7zHJ /mx6pL5j/\^/|*ϩf#g]5ró)ZJ5nWR%ss(Ә~CD]gj!>SS PE&HҺ7K((Gc&+XŒ߬DNe0f(>HZ* %L|'2AumS<6],&a*@>2NO/,d"H_|<{{v˛B7?xJKA'r\,<,/Oeܰ,?{=UmoR$YMJI֌5YBWt^K2 ,Z(f}1$/: IݺI˵s P4eD>)JD̂P&9xָtF}mߵ4r贫!IXIIQZJu\JUI2$?ۻog[|&JaHAT?K e".!r~>\'[}Zxk6I.dy ݑlI>]J%G6g0Yro/l')M:Oن7fҪ{mY>3ofPsj-'xNzN"efffaf:tttd^`3-zmڦ.ܺOgOn?~O?_ڸkv+V@]U#V ܖ+Y KmR7H/yvgn{wN$Fǻ,ڕtο= wYύm\ߍ&"VV~ɰMl`7WW9bY v59Xm+[k.q&{.5B-q7y;뢠RKqԢ}dTJ)brE8ؐȺ+$]:'D|'Î^弓 қC)Y4%RERgeZSz;OR7ط ˨~5w $ς;ɔH@Oݹ@<#Eq @!z/$ *> 8"a(% Z%Y=h̗j KK^ MJ9I*;\l.ymCz!冯t͙7NW[clܐ\O@h !r-1Z+t:%I=`Fe4 BU"q22{¿\ .:Sh|/a1Ⱦy?+n2 '2y6+rx_ԃyTFu$^@FQX%yaK ƪCaT9>42Bц,3c1NleAcF9Ʀ*-'xT`K$U>M)ѕU썜1 "?_hcV<bs*c?6>{Ɖ&h N%4%MmaDd,b*_Z[si2MޖWxg8YFi׳r[fX+(v|;7O8iÝ7,ik6?YQA3Lq?F{=7l>Vgl]_mqXy/%EgӶ8DPOM%(>\vk'GL *]A#,%@B$C,ۅ,)n m^3#{ѲpꊱZ:e3:L2FeO:9a!:k)).eju 'Ue)$KU6c 0SF6%[S{ F}ylv,g--AXe 'oG :qonL{뉱8%6U3 fOWu0YL믢>ϱcēC`M;rtRucwc~AO0v +|sAgAWɐ1RZ0Z*.VASKZH )XMvʱt峴 c2 c{#gK1$.ƣ >OË?z.j꟔6K׏ qmsNF)TR #IԾ2:r$1Qv)X57ejy)[R6@dS1Tbmԕe)״^9K8n ZY<{UOlP!7.y^1hOYawiXI<>)>IK~,t тr|qG6֐*Ep"y#!_ *Q˓~}[oyPj189;ct{ƣtTOK?٧S~|8%>a)Ke9fLUUL|J^Oe hHB2e.ime(GMXѽ\){ɫPJ&PrDA9+1M!uT9[{ 0~>! aO1R\|=_mgM櫶!.l2ͪWeZV4bo͐٠옽)}ףˮOZ$_ٳIs؀]^?[&0Zpn{-.l>=g[ n)fܺ}7M=_5KEK-ofOdWcކGuS u~K3sZbF}Ck·?om 1J(Wm:?Ǹǒ6O;單;ZBăO3}ʳz>b-o_Y.GpC̰WSXupUDhE9OTZ.mC͌t BWx[U.dԔٻFn$W 6=Co\7\6"'NcNf+#{F,QVk܃3Vd*[݃sقzyKu}&:qX'77K$5zfvmQ}f.9]>z'Ƭ,53JJh{Jc31R|N>hE'h-X.tM:2ϓĿ3UOѯ+ӳ>f:=ڥ|W^_fywaSc&_M{, }Y{ j$m.hkl!fIu`Tִ (V{@KPp2d--06p7^j/^BwlOnNJaq~ݨUj.\Q K0]M)3Q̉~.&qczwr:;* \)l&# )5 82R8cK,T pXx/N:mW]0`~鿟 ]8?<3b{ QrfVt^YV׬P2r^P8VQlmi_ %TMFd#A8_ޔHֵnbߊ;^iɸ j7mQ6Fm3-n E̾T_ABJYyYn)XQs5_J66DY3Vtlk(dU)("M"Qmh]ܠ9Ԩ_Dl}AD;  k9Cmf2"()i@[8 Ea2ZV!@EdR,m5X dR8Ɇi.%@HڳuVlЋK'>u6-q1.\\`Hj# 贈 #g[RP˶a"$dZ \ 6㎞|' aQ,Ee]Tܐri_z.kW|B|w:9 slw_ s [Ԡpɰ-CXf#f x X#R-RAiYBl9%_" jjN(*m=DP$,XMj Xå(33 F+r:dRuVl?Y49;oLmb* &ALoMl!rts)]"I;H}lS5t~W[ ,\eHL;Vk Q X-Ndi,m=M]}|C2uz_P DQz\**F+kl9+l=S_m*:kzNy _3gl=5?V>@*f +1"McJ.McJR߹6FT74F|[+ uKe8tKP kr>/GOd&Ft&K.?GIX*cbCS٤ 럨OУ=|~ú~AB(J+[AAa"1;r֠Rd= ,Eg}0bzAKqR]&msL6#2̵0{BO t销AI}cO~*U2BqcD=I)jmٝ ?FR ;x>H#]ѣ*P(Z"c=$c-8glzSyMe_7-/I /ɬ cUN/!gH?=Lw=S&XKs4駕\ǒ~Z5駕J&Top5p$QEsXz@O~~\]zU%U%c+%*K]z9p$Kq>;bISOXYfLxfQ`]VwxOS|(a93(AH~og%Tr4LYpaן)S÷_٨5N+ K1be7XMy:^sLj^s4j4Wj^ҙA{j^wLzQ\ }_؜OltՒcߘ!GS[ Fg?rtuq=x6yqѝ4#LOUJ9m?|{o;C)cWr5 LWj]xuwT. 0r`wKW`wҖxWh|46I킇8a;=q3BV]D;1waaX=&cy)1c:|p i-KmF bD2?6!uC"[x|[vK%5VEJWW]fky಴MwMԤEA5W<OxrAbmlS+l"ATuva-N+鄣 LfpxBEK SZrlC'ȩ gXy.ٰ&OQlRTJRǛ˴U+rI> _/ IހK"qRb٧ d#/sCb(tMDGd6F(2S!|0.$#Nbq4#2]ᮺa铰EF%o*Jku>y,Ul=XD0)`c]DYRCXضWX-k,4NsIfTr`#$JImNJ%\VW IZ65QؼCԿ-տu&x7-usys1Vl#wgszÕ9liN!w]v~bvӦsmt'KwmqAvMx5` I)H6yUK:3}_434ڀey&_;[YddQj.9%֍MwI]Ӡmsi/ ^WaĒPEU@!ZlTY9EL0X'R8;6%pif%Q۱-kw_MΖJ4ɱZi ztbH.xn5Z'+?NɰcCGN5$["I̪ʶb9M=g 1U) ZbZjd(~XEG|gNeEo4qja\w<A9n^mUCy0Ŧ,t.\Osu!n BvlmRV9u6:#,cۭdw:"joёA8GFl;6Q{X}$ٶ`͂T.f/H7q3_RBRf {MM JmsWo#kbԺ(Jt5[Hp*Vhe (%cĞqRo96 8=Mm[6ovZ5\nݢOݸkL>׸k'Ev <tL(4W㣣F>.d]х$-w&FlVaFө*?3c L|"H'D;߬HTK֘DT1hkZcK p2G!"x s\@6QվuŨQ(1_|PPztCM)}J{_|z?Mwa'X=3}G>r=yy+]F}؇!Q MgAQ]d0ԓ}̲cVΩr==q9^aI~Ii<״o+=xT#)yٷ|qv-ڙM? l۲S9'tŧo;ZmۢKNǻ[͉5e~ r&zn- jG7ۨj]$=JH\ʅh~ ԄM׿mYWk,s?viuwA,򱳜C0Z qr4h~CdF)\ZWs545e0m6٧VeH"-ܫ\["*ڟhN<MlNfpyǺ.U(|-6xsQݽ+ :ƋZGjxf$1Q#jHڕTtFrB:¢6bP{ 6#RfrlZ FA]O<7;ɃG~/{̢)'[ayJDffWÀW &pЋϹ|jnI?2>?8h3ȩKMq\ c ƚӆ` $`jA\S. Fk16T\vZ-@eJ}w1`!()tD\DrZ#5#ȜJ=ަsdź1]/~Jp6~{5>:מ(m+^cдl̊)!'S\uɘp)Rq3 7&)SKuFլ|-bĜ<'}J ,+9TQ̎W͎'ަs⬟@-_^^]Y^,jrI;T:TܣW hQ AZ>$RrdVSAY:W@S8]oyN?}P `A)9~PU'ڟH1k.nPa|t[l#!غ1b"j_lQ5QZ {)!AHh<\J".Pt޾sX]t-%)r1b @B *XŖrI8E&9+!)) [_xB}`2(ǎ[wqA)_Xe"k.e&ܻX#L14#p2L:"S|uݜ_=A&AfͺZtT刔 !Cp$S-5;!siOk{qdo-hk٤h1h^<%瞣(PdzRU]kٜ|'wZ]\kS[/>W & lF L}4퀴ifaʼnȄM/&ND J86[sj,Wq"/=NRKԗ\~;kGRRκ((r@Ys!`BԐ/h6d2:1ĬIhW*Dot@PUWWy`^nUĨ3ՀT2bֲ6UP@2,Zh|Ϯަsa&$ +;_trص9׮q呾5dz7l=p)|/ĕj^K:v$Nm *L}/Unz[~72 [jSIw7ѝF�,Mc'qjNo(#;ͩ[[w|se;O-iWn}qϾz󔜍꽏Wmx0xc=;۞ߌyʔ48=EJL:L3V4ا) IFgy4= fITHte(+1Ei`OMew:*ҿN際XrA!Z"\lʊ: `Aq c{0)2};D ipcO/ί'3@;P[.RS0̅2dlbu\SZ5d&73?aU' T7Qta$(%Kόt3*OK1 ,3b[˓Z F$]ފ|5ڲ 3ʇS"(m KB 0"C39GZlMKN{>JTUk7c ڈ* eX́IV!WR}KަsD2koX7YeSۯ,4j* ׼9_2^nvH m_ـ.FFO_Y%vJPJl*9d8[;Ę1Ol21)%Ϊ(%կĆ24e/Q L"(D2*jlb|-\U gtKlTvoq_z~[T +T47AiMM5mEJojZA@ FhȂ -5Īx# d(%e2Ru(reH{ay8U _`<}D D$^'Y)=Ckd`u)D(Zu|3[oQ_DZ,T}@A4&g+-{2GdieteQdtJިrg8Ž%'9RM)HLPn-DM,][o#+B"88A^}9b;+3 俧%%Krjmz,&.XEO[㎎և'Pa-wcW+~hN(`zcNfk1ϵ<0b !d10-iHРJ.ԇ:vmmF1D"DI(%9EGiXcR E!1`vO)I`uhi.FkJD#X7_YlQm:5DZ3kfMCm*ۙ4s#ïg@}G11k4axZ=LxmIggp>%c7XA+_P PV&,FѤ59?T[utS;|\_-ǛŸU^MN<[>RN Rd]٬@@'Ж))DZIe%~{pN xg$;MS Yy>i\4* #k8m&X$z# EV01$"L3&Kֲ `=A QAm60% wK{U~0 7{jS#}F|;5e="BOt:]h &0LKRVl1D Mla3~H]R=[K:\'I($9]-IS!^JP6RkoKoփ%So/2 <:ʸI 5"n2cy? o{v/Y\`i :6W/ŗ'Jն|z'ߣ8U #BG@![a$-c4t[kwْs7x]EeNgFƤkiO9&QIĖ^:^+Jr`,$th[ql|Ց.'IR?6O?=S) ~sc5lc/O6cAB m6yO3 S IAu/z7ٹVoСVoЍ'A'm̶ZC5%E1 O,CLp3NХnp<`<, a}z_+NXjYLUb,P[R6 5'so|{}7r*o8o[zLATYls} v҂`sɔ&`vVx[n(*tT=â402^d\3;m]jg^ZHyY`mFY7&kŶQd)c (>XZ')kS5w]7ac( V҄ iS}]R6mBy7Lga&?/0c,Hdtͼ5p}km! Z5飹|9ziI?68hu|.n^֚e*@߆ n0xusï4D_i^?`ɺvoԥQ&`pMM '*krT>\8Dvɢ!0Uj48DcvJ{|2 ]'䋩ӵ^J:ە,WuoRk=w9`}x^oOYZҦ dNUR"M/ˎT p]^:r~u w{ս]xjyqHANhb`-aDB%IPN BT " )VR"4rQOfBgYd^HX[Ξ6Lxm*6L U`o^o+ޫ[s( I5.fFPVUjtK+Ԫ` Y" !s*ɒKzJQhS(dZ֦sL)Tۺ. K I{-ySaT;QN Y$|5TxOc K,S˘y[Kr B\R˛TJ@*%zhMZ)IZE,|96"ުrw%@d8}^w`sљip液Ϥ\!~h%^GgE"4Ո¿"̜M=$:)z&HzgD[W>$*9D/Qa P!ɩ6_}OeO3nw| bư}kt}< WXJu}0Q&PhB]҇dn7,":%Pd'qEFOޕe(yvKkhx!ld nm:K8Kr|dx4-ۢ▟Xzӭ!<6ԘaWy,jPyA/.ʗRٿָW$WA!02,Ovӹw n ۃkeP$eXKzHCK/]LJI4?BԼIˇ< 翰Oy`m~S=4K~F/-u=) ^ΉJ_۵k%Cؒ3B=rJ{ L֖ Q=ny( j`ݐ0 |cPSXKG1h,(5e48kq*`Eb! :ߥLLaښu_̙x;3}~CIEv,;I.,[(dLnh > "HskNyɼ[Ե!T;6Qj̔@2˚;Θx{}_ge3~}?9ǁxT}Owr<&ŪZ,e f(5<ldN5uƘBRs=SbƉ2x ɏY<,i_{S}>`:Oy&~cPp7tfeOK輌F:qG:8r.'v)hk'6] Dȡlw q_#oԸQޜyOz3u f}{z_"ًe[W\=l=[b4z9lRԇ)HR :P`|7%Ɗ:ܑada]>Q>E>L)6QJCڂ EE>#y6r'rTFKya"%D + ZI: ($uIMQ#a@M1 g 1fQ乛8}KO=p'.ћ}̝U'fW_E6i妞mRMv9wF vH@S;B ɖi rMߜ<$˫h |F)8-Knё(qc'4UG]uLx (E2^> 3)ͥ(33 F+b08[\~zI~}:~?r'ʻ,'sd}M9N`a~ +m%>S(Q;P<М2l.]T7 (ؕF6m#t y\^XMFA$K{NZ*GY bD2mBB2@:gذ (oM({kʾ{}y>h2@Cu\T+V(@+:'J *,zO-uqײ=L#U9&2X+\jlmvni];}n`Ƀ4k4O{%+I|ݕpe[ ۅ~>yCQywJP$&2מF.hKG8*BpVT2U$-C0M8[JV4QzrL,L 6R`1HFfWi4@,T X4s7T՞gzum^.G/sb=j":G,k`KU]VR^9/{v((6  2TH `DUa!dĂy>7%"u8;^OhbiǡmP{`i"f_b B`!%,Jd[ )v\m$/ ~66DfJd ["2+Ō27BI$&نg~c_~퀈"nx" PPb!,embFk ג>D6I"s/J LL$J&1Z&jSօQ[Gw?ȸ8]\Sq 8K@RYx_ R&PPbƶfP $ .ʹ'xzYk. VƁ.]N|;JjgoO={Ny_NwoQٌ j/ӶPrtX|1X ǘ0 pDPái00d B5 t0X!I'kmgLl0mX$)Ew ғ4'(YF1SUʐZ4g70L">!g.7N=nWuga]xD8 <~LeO(yO%c4 4 3YJu%a1gAvuNe{$xoQG zyyþ~AB(J+[AAa"1;r֠Rd= yҸZs@ p\>3rkb!hILu{,;#P :NJtJ`ͬG`OF=1>t-6G ǡ:C)5./d#W^ z<a&UD>`}QBAH!Zg,1e=x+_;#G..i}ZSitz>OFwܲ0SVwGaA^$ΎK/Li`1"-,6lXQ(d<3ŌFt&nw{[NӺ\qG-<Կ\r+-M|;bRV9 2;A.:`;pI!cVNgoR'6=;״/\Qp+~(١WD銎WrDSs̜11RAzf1 rMTuo%R=K/=V9TnO:]#J5q~ϐ~L{Zq"c1{qA9J"Cm}$,R`:N]#]Q#]G$M>daQ"XCYCQ!Ԅ'G]Dp >uͅƟKBxYu6mѣvPLTXXpl7NfN٨59twyZ.`̯(t_uq$8n-w}&թu^~yV}iXJ`sΔ.*r`'g[ μmz+}ow#Fݧ jq#pXrʗO ep-U^Ka=wٙ҂_uMG/c`EPg̼y Un%7^9.yMl3_mZruof}Tj~"uBF2KoP)Tv0,E=g ΰ |C FWQɰVrk!t:xh9fmR4iI1]%;WwƪdUBITOWv`*@ N()ҕPwj' u$UBIiOW|^XwǺUBKT*dWHWJ#->WЮUBp*ԽuJk; `Etg jޙ}Z e PʞNWЫ+tU B>VP.U=BtU%o3jЕءO7ϡ}.d<\a9-fQ!?>oål5ln%,@2F9Af ɿQg/MHVoNj/՜uNèz*xD;;B]yWIWݽM&p;SEj v"N*NδQ*<)elvWʹs|nϴ\%ݩdbv,N ;ܔ,7%3M snܔPj/7&J<3L2ϐ*oR!S;7h7nj4u\FB mR_!M3W ]`| Jhm;]JRctzs1]%Bw\늃ՂZo]%&tzJ$: &;the*ĬWHWRPx*ּ3t%CwZBNW eNctUJ AH" PWu-E&Xt J !Jew" 3 BOW %ڿ^_!)Y tUzp兓GC/-t{:vBU|Ճ{igZZOW eۖ{*t>VOkbx(o(`FVv_#ՠEϞQ3c,;c%_'y -m7J!z3yhA[{ g \)Ԙ[v+9M`Igh:h:e4P_#M3U a{l-~Uׄ.YW]* N(ҕ Ru5|SqS3'97peqݛ{3Ӊ/'܏M[~?\NI\22ꨔ HƜDgHgcu~.l25/RB_yO}2ܬ4qy/?Uh6}e/'aFXim,虲aČA. =%32LS2KtERe_'@fO} j]B1vhݿ./cሸҀ|S#* `/j] 0o\^R2edp7i1kYpb2Se@_̒*RuI7H?@2pȘESQFHL|~.ko08 A1.feU~SB˷|ocЌH^ᮛ7-~r w{~[Ai}7ؐn4ztn-btyϦgwWGWS%lBЖw~ ey+Q_Q GCAM)2 . " ؂-+p?mvgӪf0y|ivϮP..nd&#QHYuZ30{-#ce46BZ")p\qYK2x&^.*Gfr4_*3d>-;\X=Asَef $--Sx<80O84ƹu3vtS)7ߤדxs˷}-zHo2X+,WZ*w8̪n\6[SQU4 #TdҒ༡@T  A"X׀[Uka 9Ny`p$PNbXX)r +jcp+jKfD=m Hf]-^]zi<nْlt2E%Z jRYRwEOru@Mi]a2_ʶ?v6aI(\8k@ɑ~pKa6cia$R96F4ʌB裵(}C(#$N9ťz xe1h] Lt,`^ۇ`o'OmWRNjKɂn<.W7Z26g FTb4/n0> poʗ0ïl1,f4-/m3&B,"tO*Rݏ6'g w\ϖO2 X<'=S!O˘Ao a͛ m$T{6 eK Gm2:J лk qt/DASE1U:`\d"QCءa9$$V8m|~CL1> :tE N&x6S!?KΦ=5Yi{WOB2tRaSp+$+`p448J)q:Π122(,"&" 3=xcE4Fu (WKM)0jZRʃp-l^jph4M; kJ{4TրwLpSmTUд6EW0pV4MxyIBB]*dXőp.uXE0 ZIPI&r4?@By^傑uIǨD2xZ0Iǥes!) ܃q6:`jpc-ұYp)`VOӺv+],Н,g1_r-=y%Jn@KbL_\e ٰmQh0\? R/CazJ)K-hW:s#g5*Cz Y@SUknNo+$~-mPʕ&w#s $~pƃKS>٫9 ³ϧ]CO;B [lbŬ_ؕz> ]mnKm45- BJrv[[&=֥,ߡ\xz;\ /J\x0^>gΒu+} yɢ7>c\M(bxDΔ ;FmrTiQXOe)I#4CkG!em x0Êp?F#Ot:­IPPי4{,\vo@6O!dSUZ?]L. JgP:o*-KszB MuyHӧCJk`mwlG6~W55)) A`J!cB2Ä 85+V@ P\;9ڪ?ߔO0-̈onW|^<hb{k9D/2& *Hc)BZ3-ЗKTd`2Y{pPhyéz /O R|Slv"|X.g|\5' jJۈ6g5$%GZ }d "NFu$${~I'~I%kЪf4jƧ $1Ŭg-Rp4< y8AIȀ>hYԵn1 #r!X)EK;f]wa?=k׾'iazPT5~(Fj`i!Ĥ`CKZ>[{` moS2:i~Ju Yoq{m.;8oy㺨̈r%\SRhy:-qq1IwU1xz]:+a3g߻dۮ#]2("˴l!}r-!-bH&dCΪu*5בSNо ApU&2nbXΖl2U^ںY?k@k.GsY#~Fʍ5AdEJgpCfhV:,y^vHOszܛXh0RL^[c$Rq%x:gU chH9c+KX߮pp72N>\=h ǃ[jW/'$Cwi粏~շ-bm>8|`?ړݧ p Y"a@V-O]Ygȓ_czowŠrٺ^*3Iz!.H呼iLNvCzYfs!_M]_K[]O9M)`i%xgӯ$F|5Rhb7.З\W5yq"nsɒN3%K:/gH2-J#hw4I[ie'"9T1\ V!8U" ƂpAK 8򐸎.r@g2B]H+Bd`i0+cTyrC%k9ٸ_gLZ]jO,!G_~K=U(%Zmн7AL^z-%sOR}jY.ɓ=n.G MNP] ` ,Rvz{&+ ca`lr~r ˌ9 5ΈJd!e.fr^sr)kSƬ0ڑ}-"O-uoN|gY/1̓mZ⚤okA}Huxĥ҃uQLdG|LiC^Ttq :΍s:T?=xi"=V :wպ8?t:6wW_oZ\D ٢Su=zEYfl>zi/48[\,*6(jn758]L=&3tB*NG\%r|w ZyU^\BqEDa ӼSTN{V(klE@U#3@l< mټc?<_aާw|)2e emˊٳ1q!_)(>:n]?{f̻oK(' TL*y*PUv*&*qپFHd q@0gd8qEն*Qh/^;(僫efᜤ<#zQzs|MǷUS9ɩ9&/KFΙBIN޹ߌ D4:1b:QpՋ#9HW@0fdU"WSWZy%Z*^\G\ I)@'r>q*Q)t/^bD4W>ˀ:0i.ȝW~RW(hH&{B+3gc-94X}S^sx|ΘOVca\`eN$ئrsSȝWM'pjtBnU2Դ2i0s@\;0I(>?ʗWmQi6o`҇PnjUnOfg.9\Og g"nx`h([=gmf~bE&(TӡE3U~ay]3łh2[>v|,&'Bf.o<폟@3,R깲URamDM0*rdCLXoh]tF~}7-;pt F7;F$蠘&B .r8S(r"VJ@oadLā.^E)K=nk*n7LhcS1&k﹙e:1N`{qaeUE)J0Ӿ͚z4 PT0qHΘOfr@f _Z>6XLy,gj`4=G }L;}LqעwEDKѸ2tAF18IJ%>!ƤP\%@PF漏;D]tJwۅ-oV'';Bv4)hjja`#oD*(M~+B䰨"rۚhw ǼFu_b~EV^Szz]_HwR.jKwl㇝f<'hAS%ƒ8#*CΔ2AKd^S=&1UUeUu$QRH8Y'?.֮[sx?nq/-㦩|f-wqgY[{=]- >ZYK0X`"k"QKàQE57V"FMsYmjIrq:2RC+5-uWa^Fc[Wa4!) oAm.%ͩ*gȈ\aasN<ŠՖ%O:ޗtm'Zp欣X«Q+9 i4E)֨(9ETnjQmGs+:2,"1rbrX+?B"۶ͷF $t$~ĖΚ"O5866|z l]aͧࣇ/1IC:Ĝ-WeX!rϕM8+ `pP-eLۑB1Ha(}]+ǣY?:lԱPu"%|`q Dq䱋kSVF1#ʴd$sE؄ A2 ǨU1( 5ńk R52Jn s> Dfi"/E'J.KwS,z}LW,[.=YYô,=;F^c4%8K&uhƠ~b2Z!e IBx̚H/_u`HBOg,#ށFǒsfEA3Jb,aE !P'aH&S;Aܶd(`0 U.f+d >b݄epJ˜C |CXlQM*R:1_#oMښ6YH^)/DN#"!ay#U_H.]\\;@)"uR{ahEpzA6((BL(W7tEY?˫]33sf:+ Bj?*i^ K@e ӫǗavݙ|r?T'ZW׿IFh |3zl6 i(.RzZN6ʅ~lvO U3ü/_rRRf'JPGL)F)1wѕryQ `H{u(^h!0?&[c&ކ8ݟk㥧{{s} ?:{2'qDI0 'dSNbH-8 ՜DRGTV:hanUˋۛGV)4M} ؙa/yWcxzM.*B3F+gzq% 3g 3M@tϧ:AR~ħHsΟN۟\JvU>Րl8Oq9^ֈm3k;@#biY5vfJh?.ϲ|ݍ6.wl[:ޚw&Eu|&nں'9l.uQ! UN [Ju;ǽ5ɀ|!+s06 FEI |tɰ3KeNhBh :a( `T:Z5]T!~6o}vYҖ;-RlBv=As9Q \ջLnbu"i0W^ Kegn<-kOTjh9醳Y\ o" E?܌#XAT@P~*u!yXjg>HQkcuQR&1|KfZr\"g#xҟ:v=J_( m m0$kVhtݩmy>,|o9vz$W֝UU*]F&jv5ۻ騨% {l:ϒa1ըy 1 ؜t|ryk2ϴ:Cfrlhs~}sX̵Q*g er+gI1cV8nzE{ݳ,32s$(k"J MH:HԈ:bt]rתZ5Sb+t@HmojA;[,A2Vٖպ<{&:7T3־5uXRZș YUرMS馽\Ӎ[(o{e9fGGmb"7CPp?9-r͙A6ƒEnChGHDԑ ⋎{>$: 9-@S-+4hOT^֣vXR+TإuPP ՛`}^AfZH_A6UYR\;{|sLԓ"8%@d&0QՕUfV>)yډw(ٳAm,mIh(;۸Zo)Ȗe>`j`W5h^]¯ Y]+ [U $XaXŦ5j >(2A %ޠU՟H6pN"fm@v1c0P&. QtlQF/T;Bh/&IyMldhR(Vw, s)=MA L<o*59ku.:*Y$6(xOAX(,!,l6,McV oRT+)A2 e-)'IJE,<22Tj]MuE]:akg ҷWVھ={3cD9d/u23j)IQz~5"p% eC(R㌢&Z+NW)eG[69^5%Bwg;ɭ h˿q]'82M-<ɷrr}ʭ6p65w]K\_.?]ξ]P/*?R}RZ*^$ M-JQnjK)[.'[[RFlil*խEyëT\!cvc eM@ B[vj+28 /efMH%(}MڲiSe*tcYѣzIw`訒*M* EᚠkT2Dd;:g=tJC#(gsd0yeB'LDA2] D%W=t,i#?я^GywF=o'?*?Uq~s~_<٫϶ÏމT=B#we[[\,ߗ=ou@ 24$7`l"P=S7lw7lѻLG} ]ɨ02h0"BA&)9- 9HPNh+">#]xX@2*6;] HTɻȓJ! =y9i,;#瀢6K\SG0yk>w nޯYj,~tdž4^; ̦ZSz՟h R+ \j%׊K:{TRnF˥rVY s37u\O9f?uQ,0oF :6U.2giWz{ [hT%`sq9E2i,"%b*Iy &&\('\$C-[$ )~KA!e_0մ%֮d<-fIhw/uluwF΁c[d}?! W+"OåZ[1 lټE}~X:Uj/}p5.*0-%HBmW 쀭W+U#yzWBPldcTQ3!xs%3rY6ͮˮ7+_ou;<^;lǮa'rƯTk\F8̫Ӷr%YG^ҳ E۲(&2#&/Q0Y XWa/1hrCV&/CR;uŅkVX^1״z?Q{+/S֓{+p;s,ۙ{93awZ>a2 x=7wVn̲Oty+'&nة`!c}w䣖[:bxO߻ğM}V=Fc(I5hKr*M6KZ-y_&?/}R[rv5l mY#j]IJZ2%P%-+Y?]Šɸ"ږ!V\Sij/)J"3:^ULmF;XbN>4:yۍE/t>ۛheʴa fO..'SLoJ!e/ RUEQo4*cTV:ڑ"jR)+邔ZTmb„%"Qvzar9k ,S QYDJ` 4yV ͝s_{[ mGˋ&Z֧prXڎ[&$iGuTb,ݎY6Ы]gWa+'];Ƚ-j8v[ԺU?t]1_,Uxnvfs^.eӋk ?٬a}{#o~xU#o=wl|.ae7˟>0yenzOi\O%wTIY<'xB'Kth,6?m^c\zfܧ^4A^}ځnSȄY6 he[^LvɎԻG-U/L_mtLfTHg/uG2V"k5 F錎޵ɾ{d#u;~߆;>Ӓ-# U{bA'E(sL=,mQ9j%@Hs$cJ mnSt`bV"]A[b3#Q:0#lXP;#aA6%շ$Zbb2(1sb_3 ;: 1 s7F?(QZxʪhTϦ' ]R슜_j-j6߷7̮EQjWJ."4",f_qW>aI?~Ş,zfCFD6>`ʔ8ֆڀ 7قFf!:CDe2| HHW0X)x׀NY*gXm86hH]s`}dH !BaAVtDM;[I,H +- BwiBkJ蘅WƆ`rJS{—DȉښZUɜ|Q zPEef22'CWa%#T_x&*((P) aGk-%2fX\0Xd}ܫ7|cҶ[p3htHֽh߲sh߽'*{_Jc_W\z8JJOWLpuʞqTZ/# Fz҃" hW\rCJ}J \I@cŴU#e,e ?nȯ %\F1yitL<m2]30[˯ex*mIPWg6Y|bRweXcÚG KQ^`YW7û0x=3BJ`dh+dQFϔ#f riʨY 6P;_L:glLCV^g/ǦN73.g>1èi&knKYWԳM]貹ۈ$@=Ύ KVst"OEh:zr5 y3UK蔼7%sߥࣁ-T(ͩ*gȈ\aasN<,m*q]?s KYZ6n:nV[[lc;PHhA #(H8H,\Άָ|4PJ$_X>~.{i]OIMWҦ{I՟PPo-]`20KqeNy9JΊE5RCc/8Jb&<8A8B1tb _ cPc<58:k0#܁6A:$JSZe XHE<)?SDcqR"x2#Ŝ(aڶõa`=&@w:{I]?Nm [[;nk&5[1]j@{>_8qa8SUNH0%ƜS!BDkJ1n"A% ǷX9.s<_Y?h()$1(%>*j! +)Ja1 kz3xT= +(qqQE@N* `7RH*1C=Q}kT?o|DDώKb]-VGCj+mͮgȖ6x".E^ ݻp}2Nl|{6^XMo`>3*G&,m;z 1*gq<ɮz{7+]uwU dqH]vpq8 WF)щH7y?at,08~H)0^8٫w%,?jqbQ[SbH˭dkhR|a>M]T?Q(j^DZL*.b-PͰ[&f6߯\F00(R|~KRp5Nuj*v1cE1&.RNQ:կWuĖK̪BUVe wwvvO@_hhyϙDŽupCT^d<˵ @˘5.;[]v< oٯ5_)e'wQхc47W/:'XEM%Xۡͳ_^j|B/2ן h;7__͠x6C!Y7fkrЙ~ w)чǏs%1'etp:*(d" T%ɄR5u}Nj$"/c+mv93*\)pC[*.JS4ҧH %z5٣`+G+F2ƗGCvـK};Kw?@d_P;{?*U]} 5s66oE̞Wě۝w^4dmmL:fՁ4Ƌ֝rULq7gߕg F?R'Hv-c3_Lܿ{e_M#EH=k !ݖnE:ҋz;#1Y{I?:*'0>;T|NyT'ŁaJ_0hs5Vù֏#}.9L{wP4Z7<0HnmoxխGi#A=&YThdkje[N9l #Ⱦ^Fk<U8pu>h #l$jaV! cIGՌNV<#s_sgK-xi}I{ci"`:<" M3"!bt0 F֜^+L/aWyiؓgm 7,hc |7#} 8Sd(=:Tx}*{RjİwJjYZHS8EmG}3d`忍o?[HtmzY$jz$zJtqD]P|l.vZ% dUc#]ѧb8P\HӁSg)Nʐ3%bnyb=~rMuMOcVUG{JƓe7V5[\az ||G;-Lw^-{5]- É"uvbqgՋ**cTR1[p΂VNE0H kJNg 1w%#|mO{&véJnb({]dJ-IiY$R`,"I9OrE؄ A2 ǨU1kQRk ,Ђ|l&Z6ΆÔ]ulm#Ah؅E X2|Edj&7PXkj%@(:vDo,DFI{n~DaUV)4pP2pK h cJWq'JL -;'b9&bjba8TLANi pz)!#$EpP!GщXA,Kǘl5T9 Tc0F'si#AJ5cI̠d@4rv,M۟z|fIE *FKUq Ɋ#]4^`rXxdeZ\ڞYr'K˧Q _- = [o( D)y # r\J%r4IHO9I`9\S{ נTxG`p"%OZHg^afXTqXE${pce0H4' Rbe:pJh$A (8U it.,9gA%](EPQQPi$]1"P8QZXH N.!b WN:K{(pG+ A)GbRDG88=bʯ;^?4ٓKз&տhYX+yѻ/,Rf^fW0SnX?F|VHNXje,?2q\`l)f)QkB~pedV2k0UgՎd/Υ7(7%nntJYfvSAyF#fr{썮oōOWNk80Y3F Vc_5A"c!lL5d"HmpqI~mxk@O6,bG`Xl@6$|\I$%Ig*CgDuRv;7\& IB&ޅ{?̯ۻzuVn= Uf(-..+v7 ;M@tO:܍O??@˶%Y8Ymi3l]dmse+w7baYNU 2 1r_e _._ 2wقضp5L狠R'^YVePr+E҆T1:%fbo)Ip'vc-q].:A¦R ?ɰC]Gէ [ D"lENSJGi)&FUA[;k쨒ƵC 7h_lal\urS>yQOHTl8Q6=QtSl7 X8I Mg#CIِ<5[DhdJTbd)[ERyŹy l[G7_c}~9b}XC}|]_['yswO^ɗ'3]MpO~_(\ j-`m'ޭ@^RQn6~:};@jsS4>NT3N6O SMNTpGmboZkTU5A2Zdd65:kѺ\1QA'uhG3 B/y4lwb, `߼}{6FWql$ˏeQ/mO`V) Sj&vfBCC}gz-s\(xA! *rk5D %R69GոwQ\ߓntb{kNgOOjֹ;Ջ9T`ơr.K-,.m,729dYn1Y@n.8b/{tqyVRd\ЈmVt8yWa"CyJ ˤqlhL~j]G3},BEoM/.`J$<ʺCъ5Zqj&Kt92T`WRXoBp1gʺ[Sc|1LRGG K]ϝr5k_o3iRVGe |k>g$sK߷ ^'1udbJ#žYM%}2}:w2"urp\DmqYi)ҁ>McW>XG45]+̱TD7C6k)TPi0iQ4%˽:> ) !PBBJʗB@LLccse:;?>'4ҝo+!~b+B\6Euߗ׬ټ$B_/_.5>̩*C*.q66UCf( ! K.1@fVll)8rN&R/s%ǜ5bQjuū,m:hE͜n#WR{7^}36,=(%JY|V+$WcTjUsPt:DJtDr8ۉUc?>Yb!3JQbOAUP#x{~g= ]5e_RgȖ]A q*|aqktw!QjSF-ImE0fWB[hCϽM?2\Eu ݫlܜ@r^W{_c N %عҌ1&VlንT3Zz~2/ܱ7uݤh3|%w ŀ<}_O zCȉrmurb!]Vzu(xH5awFu ,?X^iy5?ߖL}8X"|dEm`'4 9dg{{<Od,\ŘTg[MD@ ot(5vS5Dj&@U͈@%XeJ|`= gGAEy?J~ɳۧ,qWomݭwuVL܏unu.jr.O.8wtB+EO[83܉w/>[Փ^zZ^}u:O.CH\Pi7o^w^~}Q~?͘wӓWx4zRZ=Gתy5Kw=z)jysKMoχ^"Fzr;d`l|CլT`fYuqѼjV]^JEpf VJfWbr]۷(\ 2ȌSA)lÈaD7#0>JFqZ.PBqLf~s ƈ.l6_##>sxS혺G'Gȅψ}CHa[LZ$cZml k|eOVJ|7d>?y+R]?;?GRuc$eB&Z?/j]ɟ>g,6Ÿ 1h6˟bA-befB[+?/yS燢 P*brI9qEn;1nٜP|&ېA~(2Hc6݌QU詥 [UbkCR,8LYqaX8L3b5,|V,u|gg3_x^v㢬*_mS>>>~|Z_SR"RbO5UN6(] ro3 %1D(4tc/5æCHM%415KRfJ{8OOdE0 ޺~FpMK_?I&h9ۚX%jײ!0nM'Ps>H5(H5[5 R)(~YSԨ!7Lx0v`DဈDWE(㽯k@@R WT(0&{=Jc6Rf]c%R%Bj>JPK JzRtZ ؗdeG%]?W}լxqhOaZH\D\R;? 8MZ!q8sfB@XNEt9s0<\x 60M#qGh@ =+k?L\G0iba_4XAR)_`m<k?+8:fઋ[.mLW*\_W>O..H"kif)g+,2 c{W2n:)_.E07wsh>nh;\튋uJ'C)vhɐ_G72%fOh*f2J˕[PíW-?WOOG'R-C1j7/jlTn Svh%Dj .YC .EzpH3ĊF)\M{uօ,ʵ P F|.}m*6M68W3'q"=Q0:M$U@:ރ!08Rt̙tB쬴Kl$?R)sթ"Ur'Jt5犰Uܜ06WUdpʤ\ \՘H=KA bجڬ r /2P15V8jQOkWaF*3%̐A*4& Wpܢ E&+f M_"ѕː|\h]UC+L똽+F_5jHA ɫ*'e֍ˆK&#Ak>R/ώslb-ux ։UYQ2qv>2S*U< P;nΊ}Ur#o%\[pLխd}ۦay pJ "lu]B#pDWM&5- ɍvD7z8t_,X F2zBC_Mbr ѕ3*brU2+Q0tf@Xj{&x RuwvR%hBQjpNȧTkX8З*JUTwDusoQSn3,%aD]8eyγ> cArF(8JVֵ &_xNX] : FDlָX:IXbaMIqKe4'hZT E\MٻFW wV~H b'ϕjI]Q]!ER+gU6Ùst!MM(`40JƎQ+a\:H'E $;&򬐛Հ\I1#h`\AS@`!ĐiY03u*nF)K<";1 :h,,u5a0 ?%RTQ6%L+, q+9nLWn_iddKxQ3Dpa7h,X \QA5VJx5=*(I}^R6Ԉ &1bqѳT+UQc%Gji"j8(c e0>j2a+"Ǜ ~\uXgLl; x3V{[[y~2mƬ*$c QIb}0@K x1* 7Db6y#e}tbQV]k)Zc2H=f n!*(S`a U0H\\΀B.ds2V53chdr[< thAMGvbhpXd5H8KMF|f?Aj؁0"ep% ~` Ü@#܁?U2ӉTIXsuicy!ui`J fB޼jPR\ ^`vk,WH6 "kӤMkTn0sɺ6Z43FiW>vw2t1/k3  @][ 覫2h&x ѹa0`]ӿ4,E4^Guk I<5jhx9q=)gor7Qj>sp0(A/7C|C 8=ޚ̠!FXm(1⌞T4/h .Z6l!&XKo&b] 12zQ"H"b á Kr -0 Y&y]p]ꊄީED8Øc "vիma^۽qܑݲ,flm ]uc;{*{/>UT_XuPz)"mit:{_↑n [EçR33%.6k%?x%T@Q o5)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@W dlJJ lJOF k[yϽ/ZΘ׫y~ٽ?@\ߴK{MGיC8M.yw1_.<R܂kix7G椓wr}΀Eo.+~oXF.(S osm-úu^\7O`nwL)|+.޲ "VwjE|lDP?v*xӁ֮mB))$%)(IAI JRP$%)(IAI JRP$%)(IAI JRP$%)(IAI JRP$%)(IAI JRP$%)(IAI JRP_Ȋfo~~hw*%/Wgiq5|HH) >OG s~OXi5 >_mhzBmvѧKaL@PK+Bⷣ^v6ۚ(ڕ& =}~{j!@ٿ036X\.eq͘:V☻F=M U*ɘ` \ɦznVIai-lB M&\5sJjWJKpeچ)`LƊ+Xk+Xwp•Rm3]vyX6ėoA'߽9."/wݶ-~Jۿ!DRwB*n@A[<~nԓb0ب鬧4sJlnڃ?@Yi=5fcS:c'`n*)ϼ>p+9*Õ31p n2᪙kTU׆1bNRznˣGlKǙ+_ue[E="\) WOznPjBd̕L%\5k9pլ+ WKg틗NY}tNhRY}vyvOr& 돺/tNGjRw{ /8TOMi)Ʈ":%YժKΉo>ct禗u-j]ѣӛ>Sh{F2Ny=!0M 0|+d_!Tl%no]mCQMR,cc7oG܊7O;P>O^۱+ qLPEk(ޮ4 ֭js+Ͻ;{*_eVu7QN;M꜍"REީmݩ*Ƚ9cc9T9nN,`{>`O@llcsMe٣+pP/=e{}%ؗsƍ׍EvH++՗\.oDtJT!e>U`άvĉ͵TU~kg_+is;FekNk_؜_uWrYwCj'}cժ/돖 0î}՛.z5m?X_/SXgWgz؇ds\-/N?쾱._zC7!.\]Y缻ݶ-KYTnTBJׯ{:\hљ>\tu!͉Sm_WӪp8J=G0KNUY>c!/ˇ}Jb٧˓ռOkw8ʗu-?~Bm䷳_%q2uo~/>w5o~"x?ڔ/槀 7=@s|ٸ@q 6V5H=(z!HeSD'[3fSYm{+JG+reIWc7rO#JJbvשdLX_:=ipUgl3]&$ _<)LƷXUԡH2HQ¨ܪ*MYRYYVLx]6V$EUɩUR9O#&"o_;|ѵ6^ >p@Mp.ݾ?3KO׭%bjWr.90[dm8̺ꋑl0ɘk)\ g5.`ߺ5]G~{)4{s {cbf阗Q7O١lDt%AǚbרM²W(FY]I9VJϋ>\0k&imlOzשּׁ5q5$|DyNslw7?kv^q͝Aܗ0R:qQԲ3N1:_e!W]}%= ?l= idmIJ]= K rJFs8[iPscoqNp4s\o`鯝h썯ayޗL\o׶Փug䏰'%M UayVwθ:X7=\=XF6kЭox4yEl56ޖR:ÔTsBI mձbmWа-smd0z>uOWkX@9ť4NTjd,eV4NE`[qvb8K _D|&X/\  32(*2pɂ-^˜v:Z"%v狳N=wuO_ً,9MhiF'v[_waۏ۷6~vY!\@:ev6 ,aƐp3F FMckmX9 W7A_%ɐl9!EETSĶ4S3SuUu]@h&6F ~5NϝEyu"0  J1rv, KJw$E2Y@P0$ K .TȌo5A#4%ToT6+P ՂHJ;OuRsi".jr%bI `V{UU..Xd Lb +"RP;!|""BD@%y]ȹO&R,MU|Ӳ;FZw:m+=78z/3yAZQ94^Ta%<$W& \ŭߩQ68W>7yŴW%3w \v};;hŠ.LEʧœ_qOu&Rbc5@U 6T\hm`+ \PHN";X&s@eJ ƃ"$DcL j&jR+b1JX451)sr)K$E MiI,Tw 6v?xX9Tkz'.{jwdjkRoae5ћx^eb 9.[nWݧL/9J2Wڞۙȍs7OF0+}oqvw8~)\ePK[y [K\ceքljہ]r^[qTޞM{#8BNpcYE.$ AZ (2JR9JCz[THƱ`c)qb@] ,,XY*Űg쉅BFz,<(|̚bsv۾7*;M>O_GPB:'#3LL*\^[ ,pô!5;N4(ʓ א=D`6L@ m&r#x #v1r#Ê.*ڲG^ Mr>?Dfhr4pS 5\SHp#( w:r5CъbMKGU -J IѨxXxX+zDq#LJmn&#⯜Rz/ "`u+ 1W5֡@ݖ0")mjTBK1 (<(ƨGd\AI{[zDEqq쒯쉋qQKZ\!P!i ɘl y!mpsfh&!$  =.Ÿ#x`<<FyTT >Xy%яaY mCP?kdz;.Hp}ă"8LzI>佌d~$[Bڇ>9  ֍C.q0y1vTKUn4]bԒu$kRC4P]"C̟3G$^Uҧm  ΆWRwkx`ڢ0zk^Nw( DŽ݄1ca)e-Sa,:hhzZK4%EX2@NEj8X{gy=yǂ;շY(¡@wV3KƾxXG ;U].eA7c(8SB(҄E-Z'e5qG̖s_}Ͽ(t!\tDJKmxH.0 ]t,\YPYFU5GB!F*5V c<¬UtP FMr3}!Ҽ-W}{z)ٟٗAbZj((,F9DKEF jrB5GcoTh6aEE.۩v}&U\]*ː+d\Mżʐ"bŵL(b R}rKuC4U˦/J;FƋ~&|v@l|KO6ǎ#qcFYSZ%8piJFx0LJ0DŽh#SM.sRgաߏ$/;A7v1j`Eބy. ӆ_81D?iM&]n41x7ޯ!o^4˜㓳QS~x}|3D/gq O8j8, UqT>墦LL.uRzEҶ{g:iW[*+kߏqJ'z͂yz㹥SjlX$,d&WWA3"C-qxlcqB0;.1\)3&{,  Py_G:oO/mϋ-AM ^AMFRM5͟k۝õ SԆy`H:G(ȓ9GH䲓9GH sD"9 `\;tW:@[/EIyI[dU;rgQd(m+u}?%$/aK@ٻ?ބ_~X̺u}UlreZe'oo4|?}|xY*Q?,;R R/濏RHXC>SRA9KÄ"Wu*W  17t, ׻WASUSǻϣwֲ]!YЌJN/sXi^7[rJXM5:lqht;*bJ|6ָ0mݕJ/X# ?>o-س;\f _53kT*Z.|_9ݵ/'ryα|08-ZoMsWvZL/|) l!HSkF-Rjqo w2`elWF9qMSiA )W(5yX:倭1T)b+t@Hm_ja ;S4ecmNjXՀԢLŕ0|Pc2I 8+q0>u·x\[Ζvr쬣籎T:|?%f,b6L"/aN(Esfͭ$G~[-7:hO'2;zhȚ7b`%`ᴱQHΜp4OTʁdGb<Ղ8 GsXGEdQ0A['RHDc5#q> }UXi/B+}Փ`g< P}JQ ᓉMr|*QZlJ)(SbrBp~|m XOupߒ Ivڅ`yfڍ\Jv_nT .;j_cċC ƈ \%rUU^ \L3)*؞MѧΤnl(Ex?faX{r8nQdzt͇\Yk;h0y0!~?=$ v@S/6b0 RT} h*~:%$-_;j:<hkc*aLYYݪpm*5 OKk~2fK?3/QKxͼD%坙<*,Og-NZZWJ,:zpňzKn ػ:TWӽ͒I=R{0h-%>xMF9T=2CG$,-τRDiJ:N꓁in;LU!:%4XrBp:Hby*ppUJh}olUv碥YG1 «Q+9 i4E)֨(9<0BR!4R6:-J(0Dn=k8>iO~i.(f.S?4>2^3^`bKqst;̗,fXh]#206x!rə1זpSd#&mחX[dcv i҆^֟E%gKu/X?{Jy[5VjC!GKI4F$ I4uǔׄ 0#BXYL^jLDL ` Xy$RD6#g"k\>>}.&6{۫cз޺AnjC?>tbfbz_TR+ <^2,k; 3:HkP ^pʕL-y(d6@0`Pc<k&pNupC9aF"칃騒:yJ;%ÒE*/2DOy"k/%%?XZYh6}٢^M@[pꧥJ{S㪚V?h1ISU5]3-&oxfTm{L1'/'aFXi0j+5)gR3\V(:>a(zj3ScrMdt]gW7̍_hi:Řq4yT:RLy4n6|."}b~.Pu,8 E6 q=EٯQ*[|SA/oXԐr!OT hw:_N>j?q(aU9Az$xH2^P؍J, - 40KRp5 ף[3XG{>}B}, \s},=$ u3fb=sI{:7"qs1c9\ino\d!pez6oH^폧=h:D^x3?+\/4ѠW_WY8Mf=z^OC?a[y~hoxphHVfbepC-MLșA% @Rť_h+CgJbND28ˠ u}S,Tȷ WW v~sZLLR"vښ˙^:q#{X/(JS4?ӧt`$ ތ{`=5= ~苩/]JWɗ?:Rʤؙ>m? ]þ^]Tˋ%tX6D,.;mj[fց<,d&MG`rBZ*ux/|܍O ǿTRo?`P9|5rBptNQ7K`NLkBA__OE4dY?M1|ożz]͑M}[ZXoq4 HyWTo@ G) ˊ$TjpOf#ϼx#`05X#@Ke^V~6%wv`;W^a ?EFva뱪ͩ,TJ)Ŋ.$/U=rRj20f \O4`1.I.s daɑ"Owo `;$9w2XȢC^ɞW{%ܲvXXnV"~.tω.NolO;ɲV54YFȗ\9daPóD6ꄚ?Q1VȞ t;l6Z{Y-_7.kRAsVf oI)1Rڃh-Gs 35_U1GuP½S}Ӕz8e])v ›MidH^v {]w޲eϑuioʟ;ŸdlkN_^{trQlz j" erP\BeO3R0:lM%,Ő".䅫s5h}}{P33*[N`{㟏'mWz/??JZ!+8);%E+<]ZjgZg}mYn.%?x>ؠM:?u=ܙ1{㢾ǖrs g\9\k.~;::>_`1P0jr>M|OhjRj/=,F a{&OXv`۫;@ ۤR=Ƭ9v|]>6Tsh+mM8&у4Ʀ^JA qwopYbBGtN37$bo6(Kmh Ix %B G)A%] Σ))Љ8P] ;: fHATX?OFV,* z!8P,C֐v3N*C`OwngC<;$1;xWyeY|پk7)65 9o?s$|~̦44#~_Mz#w,橣~OW4yyh#puXeUH̃ѿ?ෟׯrx@)?qeϿ0&YS+ ?_qe?Fdӿ/<ӗaa9$;<ZJr_HX7c.5߮3|dQ}/]خ}ו:M,(Lp!duɿ5]W.nIVkN 꾚d$7k Mj7`㰦DZtxxpyvqQ +3ǻtNCӓo&f_fiuN|Z77ֆ>l=Q 2C(@ DiALE B D|C鬓Y(;Ӊ-aױ`հys.v:j1( Ac[Rce`ђ1 1F bXjaA} ߏ[ٵˍXO=/@#@}A5 !R% N)G0ZmIb@$ f"Q% ! :'.c3J1NL,&QΕ-Mjz-o'ab' J[wjJ?9БvVA/::3m)F~eoI5 @7& l0jDYiD;B"OVo$hu a=EQk&9ZMYzx 1n;KI3ˢXu*HWLqؤD(E!zOy.]w&ݐhszm >ҭ>or{@A[gBA7;<us7蓏9bFЬ9Y m&rrqE~(0H5FoMTvC出%0Fq|$A^Q]VibO n4Z6vn{ llq~r>|UeiqݫInG֯yӎ.vKZF2IR=bu6iEU\0Z9F3O~ZyYg\ ,ƂX}(VV>8/gfP`6R5E]{-(ޠ2ii7BRFj@8h EݠVqR(`!e4Y4Ģj2:-*PRZHڋJ04 ω'z.:-'ۻ^|9:fCh &iY )׾“.u+lɗ@NKJuŀs%ZN ҲR]㝣W,`4m}K?S蛞 Nص|;uYpVDYbVqx)UZ}'ĬR?bABL'<8t ?t|JDLoQ3{K p\j*vh; /c Q6$MiLA)*LF'2U^nGz HxG9Zu6&IC6ئK &.y(=ܷU^Ao>i;YU9 Y7kRpP~L'2ugL9k'Jz<|=h&FLM֭TDY&6`- `nP>3NolO;ɲœvJx뤵̀*29R TJop cP3BD@I)$ \1`6،I&`t鸄љ88{T;r.Y<AX黼} /?Qod1Xzhv*3J%:ǫg1}dW5-^or5|"Ē2kL^վuEZ~o~a}wmQr~;=0Ug; gBZAHP/ZrRQXZo\QXJa Gy(힫|tu$,M17K2&E eF06~ TKGGP2$% Qd*RN4 l O:mQZuZ*DA$hƜ"bƐTE|,YQ\ǥĹga67Um]Zg<.Fm 7{\pҷ: {guf wnKϽ/=l+8t~B*ڴPnk/Eq헮W=g\{ɭWc #|Ȥ? $NvƯ=3 ޒlmVɲEvq~JE}uKtd|/WW?}{~gWgO?|圯/K}>|X+W>]da4oR9l.^/۽6RZI#E&7n:Rfw׼x5 ; K:VZn+q?]xyMdX]\[yMvۤ)C4-Wm߶4|ܖWؾVnέN&&V,oGcoh'kALl[mSۦ+Q"M*9-WlKj}M5Ψ%O[ܱǟG9[VTۆH>Pe-RS@>8L 5:%I5 *Z1&ʏۂЇ.m@l[$ɐljqNnug8Wrیtys渼I9b"zlQʢsAN/2kr1FF|Ǩ| HCm"VRN60V$&,%9l#4-WK4(:n)gȵR',3&u\Q;*q>rcW81qVūK9ulE{W/V/>Œ?)>i u6mo[ҷIٱ\w+H QՋ\|ґWPw1 UMs|j\ُ-f`ԌJYHʹRz =k9tbt%]1U UW_܆Yϡe]u5~u5IW(QtFU]m4),HW l+Pքu%&+TAAI]1.b+Ż#5H`v՜ۮߦOg͢;w(ϳnUMiغ_˖GXW9o39Dgk[ՄsM*΃3zyIuc^(+)Td`OPQp*%TZrҫ*N0T$WGbt%^I5V+jGu(Y.M2ptĸ2![0"iZ.Fӂ\)`r4S:U5=EMg0 ɩbt%Kn\WBijF rJp/EWBוP U]MGWAXN\:{w~zz~&s!8 ~Iݧᐦ 5ݠ uqsҴz0g?}Jrt4[pz$&@Omx7 [Mw=#zFVBy`x׳z{@ˎf?Rw={}2|8b$/2H8 !i͏RtE9JˎdpM^:m#g:z}y^rn۵zQʹJGJmfmPu`fK>qpnvVjۗ^]UU[ִw\Գ6fQ_e\PLZ~B(j_O6p{u5 X}w܏ {>?a=j%fPPumփU@A `RtŴ.w]1VPu5A]!w:>F 8Y aþ#IHA0:V yhr0qi -<\atU(vˆt; 4#U .R4-&tjzI ҕPUPVu5,CSԕuΫfx"qsוP:<3i%,FW] ;1(U]MPWAs,)b`i .Rt%:{] #_,:T2c=WW#q~h gDt$ef\Wkz0+v] nRtu%&+Y_1ҵ:2s!Jػ #3 Y\&]j&sHGK*2[LsUȸ\}K CElܠo*  ] uJp=+ sוP"V]MPW2a6 apdI;lI}WYA1\,16~E q{>x$mJq2 #tUWf=J-FWC)bWi{|u%KqY=u=J)X=I{q9M٬ϲi[h B)a[B\ &}Ab`F+%]WBjfiY qG2!pU#-+J *K .R,-ni\9Zz:6[()`W׫Rt%+tuLQW^(IW l}H\KѕІV]MPW ]1[,ؕH"&{] 3UWԕwA)S Fw+u] muŔrRmt<aAPU1Z^WBInxEq{?enسFkoNW(1 G z.HWl}1}G5(2UW1  ŒuIAG\*z6F^^-3*2ѮPQp/%TڠrҪPC `V(Xl)uEܰmk՛J_]QB*sEGFBQ[jDC *Im1PVCBYK&iA|pRJh+DWu5A]9e*)`Հ[LgenԫDW^yC9}v JѕЂ]WBj LQ# lQա] ;1(]цYO /Ml8\j-18Rg6חF芪z4+T@5Jhs% ʎ*iBRMŝ d`S*aeZ63?ngD1uXPږ 7{E?^ %M0#4Ʃtot%TLThMhBi}uI{ ҕ;Sw߇ *+D6UP\,9BuuB0kWAmnCbIC Ѵi?dJjz kx&m짋\Jog?HDپ uqy޿,p{wNPmZ\*)Y÷{7BۇWJg5\ßdVjNNiO]"O/dOtwٜ̯/^,ko>v{ VX ὚;uͻ y}=cJzvf[Vۋvkխg̚\orjx6߾|)G>Uo_R^AeܬXfXEeM=̳+5-:?)-2')'tՑg=-?*͝AK&h<ߝwP[~_W4 A{5Dh|z-6hKMВΚl iNywu_Hȿ,K=?/S\ _sNg,_npucoS*@occnTC'[h! &MJvm>.`۴|"C!4jaAqMF"aO|a;mmGw=l0(P?8ݓb.*rPdmMMZp-bin4?\Km,ۄ]ҺWwJsqf:r2bp=P&PLk]v--?ZO![OgR+K}oL[&]or:bo##)&nR}\'h& s̜=.c }l9Ӡ;ݩCߵM^L.Z;)I2DBJAq-k삨H#qiȥCaж=gFq$ 8}%hU]]4[)~T>R#i 57)Bf;uF5`keB`T^# Lŭ#)Mٰ֞1*T*JvdZ2ɷ LIW!V`; n!\"U0VjVa   ʄWs{  QA\R}2ɮ3_%Ce:|=Ʋd2inB AvEm%˲fH57]\?7(c)(|k(P,Lh=4v#]7cEfJWukʃ AŘI cG !.A f|)*)ԙ5TDš#3إ~.-DACmj((Sќ E0GRF U BI[Rl ejA!NMFAꕶcu2R".5=() ElD 1r 9 :W@B}M6y-d+-%VSMȲZPҰ*4]AՊXZI །A#-ڛV̈KQE;fD$' GcE( / Uin2Z@5t1x] ި]-^ƺ L}6=DF: xs:p~>m:@CV+&{H\!iZTUFB1L: ANZL %tF\8gP4I"i Qk2!`(`GMtx L}h$kIu 5<o 7l ,:?jPJ4%wV4D4cҠNn )қDyE aLO(ƈMnƢb$ӽw:H)-и'GJFdUA?FтΨ19MAk.YJ@ZXf=ؤ=Cɗ Ug&dd, 6#TK.dP?uN^oY%B]?joZ 05^%pgV~E D + (|t* (AFW6'zʤ蹐fhJ7a#Ƚ5>dN֞ +JOQ!,iJ M\ѓFn5E!~ކ`ݤ1jEr7 "b9\Y 5I/, I负h9멓H_PFU~B:COW4"坩y0B.8! R GT/tїsqrL熉!f|@~0M!#"=qū9Pǝ'z*+ Q^ ƒZ&;_sC<]l$fɲKh$sK/W'K:^Q> MC08?y\qL**<gti˫RHjpo嬿u"nn ׶]][Iqd)Bo2?H乑?}Ii3Fp;'N @ @b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; :ƇQx@8ڃwz3 ; 'g'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v: >&'{a Dp<'5Н@v=F' @Wb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v}N P9frN M7Ï 1H;0%;@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; @N ~>yウUuj3,0X<&uQ1.Gc\Z#޸ΰq1~~|~DtE=^7]Ze2]= E;" x芌: ]P:1Uttw GCWWl NWr~ ձӕĮ7KvN~az\ՃPz0]ޮ*%🰨TtEp=Z-աTҕ:Luͩߵd5/ߎ1x >^-)j4wq]}r@a3}gafC٘^Ëʽx@LfNDsnZj!oV=[[v盓ի>m^6B0P?>o yG&MLq6ǡ+AbT-r7 V@vD (YgR[3]%5}5/q\޴z"+9ݲv<}vpT҅S$S)oNkfwf{싗iCG=vY?C;'Qi/<Ѓiѿo~5NNbvyHLgrKŰS^8>R͢Oչ!ɀ>Ĭ⠬Yr.6M DN~_\-4 '?,o VdӋkj5$#U1iQ '5H%j+&D?TT\OᗛP\O~q=FeNoߍIu}z|gzb}~M rVVGߤYO}>U-ͽ|hP%|Ɣt꦳4.Eo;30> bPCSr剥JZmt&mI^j bjvK2)1LJ a28"|H 𛳏 o7+(:_}h#w`C}?"|;lOpE-칞]^j[}Sȩ1✭55) oZʹD!EP͕#&e=$QE=<_@뷘]M$h:ԒeLx{gnٻ,E6jfiz*sƈ; !/G?mhH6U1oF!FK_I/e Jh4́ q/-OSCz(bGjH/MA4;}x|1U[4sdNvzHS{;C<8;>[V/ssW|}~ןA:)p˛/Vڟ3ĩԞoFMb_/r~~]lqVߜl_lEXΐiw|g(e1wmmJq0,lv0س3'dE[$e[nm"YEW ^Է." eEyꅻE_g2 n|\E *K߿NV$}UI 0XG TTчz(FiAURknI}yY=s|Xz=O%q~5n\)?5VSAzܥ XU޷;AGmEi;8Uرx>i#FKenm+\M"~⏭~M37?e:7o9z~qM6[@:slNRӆ:Tc`ڝn.;(jJ뻟&;p=]0slM&ğOH< ׼a&Ox' i^W(pzx7/`ִ ]X3k;֚f;'⋣Oh82gOZb*|)A k,'cxQjxK9H.=2Ksp?EJ`$BW#fjQp]b2OvU)Q(rXmyZWahu@`S jq`j{SBR^\vu1jRƆ TDՙx!J`>*HuGvPyhd/~\FםmS4-¸Iֺ b/Ş^-킏#XԎ\\FZ{C'弡RST$]䅳s9`>F52[Pߡ W|uWf1)Xʣ˫ºs8.iFx:1('h­D:wǽK`s_>GF=gݟ*c F' \y*j%z c%6S[fsz]k`-wW9U+h&8>Ϥ(Wv4&N<;jAj/Ǫ?&(&>ZFdK1dd\=cyG8"8{wXgjju;gz?jQGZ1VJ,Q`0jy%,nӜSP+cKqNN o!Hדּc P}4vDvuF= մ7[ݱtFCaR'aMJ'n*Lk \E L$u{lYraAX1\$$|BPnT)srN)TPQrpt4SfSFoO2ED}AFd^>;X.aQn3T;3a sFw^X8͎?vd $eE @eqL渗a?渽hy8S"DT0X֎ScABBvBc&STH4cݙ㖫`axtgⰖiY{MAF !ZɐurP3FrkLѩTA+dedw!sWx/,c6! ǗscTѾK` (I'XØI !imq7KiecއMp$WEd(4_TpϼYSuٗILդ.()YNqmCVύ뫳bg zs|>?=Q5 gado:/s_uM4΋>̤:c&N5{/)x^2on)*k9IQ~xNYe8I7$Iz'x)I+td׹5&jZUJ%$F,?}}J_2R2qޘ%'-:,OdY1U4H-3:KX閙V3]xT `˩.*irQT%tk6\IR|!v#o7"`!m=XeYYR[gY\.b)r3mk:ΔX'VbXa֊aĬKك(?l~PT2a|)s:E/TNX*2x'aOm9KN)Ov1WWZ8D3N 7S7,%M)YY謓^U'潴G}^Mjz(A7/C!WHg`USHI:TҪdCG䚡rͬIrʹ\3^:,4[eF` D /\*y!iny*BrnTʉy焵*}@!QX>""(b(tAPúȹFei־aJì?/ݕ|X;7= kGbY=WGG"b<%*1"t/2YA:Lp].g;P^ a6IhN QaDEy$#Pg@!4y32-RVP!;Yr3"S tf (TjN4E-r*bgCEilO䥓KH+8X7Nç/w]|rnӡNY]U]W`jvUgWE)WW^UZ'uqk Ӥhǣ+Z+ AqLW\~ͦWuӈ(p `ԾdPF|{~KFeV ΰMuc&QwKAO@ׁX h %x l:m{/V4t^D[혢;pŀ QkXW) y)([4JiA 7F1TOI)h$`$ZT.eGn2ݖ}pэ}J9:j9%0"}s6=KOKTCƱZ}XW|Y]SϾ;PRVRM,Ֆ;BDUj8~^c6Re9$dMRIǝRAq.rIQ`xNY,tQidS(|dD2E3%S g' ]W쌜Y&Φӫ4|ޮ. R8åFId, n|Ԧ9M=): eKc$r O"d2$-X %@KebE ]mW%?M8IXbאv7τoPb" ھBe/%P H0DpEQUc8- 1U_}ԣD,RYp9iP9I#Ld@@2WɽIv˼#Vz]v%{Q`S@2Z[R:`Ix2&e-nXL+PpcV{`{w_ Y}9xkmc Sŭ60o?jKM-Ǩ4TJ"q O9(ik9P۟ hydi!dBhX dɵG -#ْ޲{suO3D' ýv/_hF:Vj#9++r>X~~V7IXMyL񊵾ea ԙdJ EuvAVAI鏤H8(ƻrܝuJhsF/FMDphqY!O31h YfivRtqd$?Œ-ɲܲd ZfUb] J!ѥ74zq,a`$TGQz⌅_2 5"PLSR\i'W/cr@\\2.hL*,sɌZJR)֩xd/cG$c#J5XѲslLm;vkp 9팮?W]x!Z|Mmwij4`(9)]N{igtws5lg? M\\Yf#!-g'|O&.VsBДr f  M}WZ>tiI`JV "Q>*eE^K-Q_4dSNϞQ |cQzSKI"rH@0ϺlgOn_@2omQ5~yV~ɫo=q3ֺ.dzU[t}T96OM%^wa{ϊ=꺽g\H׶&r#iZuf6WͷwLR1Bfs^y4ȆG6Sn; ӄ-0r3?7O~3mxrXsdLzύn%/wa9lGke]nEȉvh >.Xnm~ ۼg[ v٤Vrz.L-J}麕J}kLhس+1 sNϸ˗<#i հMD0Fhz6a}lF0a܉=YBu):\cԖBbv(t(@ hY-(l=]{ר#r<Ž7u?쿟{>s&HTN4>*(&wR1M?|ܖvTF0|yy5~x:u4=;љ +-D2󕿅W @E*_R/s%IbCDچGV>B#?ߵ&f<~3Mzf3^Xe3ڗ`g_g/Wj[V*N/} `ӷkRhN$TސRHjq%:o%ɱ,Js1ZRf۾I9)p圩ۡr:УA$6&I,*N[A' 8茆49~=uwuPt)W|XPQURb-os)6={9%,A@v,p0KZR,* |ۈh[A:X,Ge^C֎R1[9FBL v.e֙BO5)45_7#olɟ_iktZL~ǟi{۸gMm&.(=Scl_Su^~˯n`(ߢRf?.>ᢷfhsKV_$Ϋ:j$7|r+23ζPK2_Ӊw<=?Vs##gK-9=]^z̭Fg>Ԓ6GpԾxa=42a&3)aCxyɽ6^Wgg 럳Y$#9WU$<@mIZXNq5'{TYaZtzzrdƣ 'o6>ҷ{m}~QϼC~ѻoZ4m XBV*Sc׬ p[b/uIGm./5$yq'CZ2y$ypqyyCEٻ"9{^臬]ߵ`ݳU֓jgX7`*Cmea׮?Wٳ~Sl z3,cdVa['_wHzRTAm,e2E^H\󥘢ZTf! &WXLr$Sfx^@Mѧǵ 5|?&Z8Pd=/@ DeAZ ѡ/gD%$"d[X?:lLן%z38#z2n&]V>:0QGGe~1}ݧƘA`Hh-L C6 9oY" l۸!pȮ]>QRY}DA2x׮ >dk=Yx)债V9m J$d"9ϼgIͶ1J c޺"~t[!QFa0W6][$Z'Dkg=$*S; du쉤L Zh yO&f0v 3F*0ˢ*EK[tZDQzd"r (cH9CZ Ѯ}eNiě1s7`1?Ϯ(}/@(h&C.^3y{"بCW~UF9PQP(3}&驽Yj< yG8O+{1 <0jb:ztuх;۳&SMQ "7$D ymٷLVw7nu^p1|}U,o7A wlU>VY2VM'0S|@X7j!`P{ {V[,@? P! Ң|2GJQ1Xg7(,64Th_MTp1;ڔP'fM( ^:`}RheQ0tN)g,kT^]Ý6T(@eH];#~78z/_hGj~7O׿7y@j2L9GPj:A6' R z4qX)z{!<ٴ3'~q=/2xs%d}s CaȹMhfM6yiڦAzVtYh:0tL!&Ȇ.N7Dej#tao ZzNȁv#%9g- !k$>L,ӥ0&*QjޏH?Z{q'j!֟o+t\Cf,]?O-b-v!8}sVBѷvmRhDZ"xHEkH6ZU> S)9ì+pShC\{qV 'm6$1Ig/J`!GAQY#&¢Jv5} g@(Igzȕ_E:!Y[șX g`@ĉ"i,Ǔd0[-_tݲ[v&Ԣխ#BR'h >p&`0^+@xM3qH[r=k^TJyzXxF9/kAЭwȢ)-p^u`3OJ/!}Gr2q t|rA|s"Bb FgmJ:E;nA'7Gm+ btLQxCUB+|* 頥 8K%nPZϑPY|EʡY}T,_@, dEٳbi N};F~3;=9_ qۋ]bxrZc"Q[Gy1fm1]\CSt` RCZjRA&I!j:H6h 1256_y)MRkAO/`qAZNT'mYπNgeSi֋ώ־^4Љ rh^10'bR;(MZwԱuWi{Zw!ui|cΛMtiLxsn@l 6ODi9y;gӷe"Lt/Y^D;gnֈP(Ӳ$X-I vcł]1݀Mx (El`=Xɥ(3ogVtȤd׼+qvUt|Y}K?%JlCʘs.$ɪ Nj~N Il)l5! o}~,?&T̊(l 55KOTHSH;I!݃lXIYm]Lҕj˜S޹‡DFl]*S($yL)mW=mm |QuM7d֩M$&ZJR)fȈ,!zdc9Ҿz~Y ޽س7-E݋'ڽ:UY>yJ DԆ o&ߥuBt*ŷ>W ? &nP@x|zdu DU/?~h*%҄Pcfۀ@-Cr!Q&0`Hh( .Zc‹X gKSݹ\Ycv˧SOggq<[lo!6bf3 >uf1O*ܡG7<:Y>=|ǤcGON#,ui3_=2JtFnׇ$yէ/yN~x~ )]- \]j[ ._y}C}r<#cmJsqs<Ђ2qXϛfPvG1#dհ+l.Fh󎹼0ȡ(;e,SvZP9m #I|؄ d: uΰaPޚ4P){w5w2b{ Qr5JxW5s%b.D )B=;]D  md/&C2Y / }FoJE$:Fٍq:5L;ڶc6jֺ~\KLD $%XLpK!Ŏ$PzMPƆ =C̢dXCY$@&vQ>"M"16t]ׯ3qvaKW+0;ӏ q@ĭOv6j ʔ4PP z\m ݖN(g }1VYfg"Ub6ٰ٥B5T5vr+qh1Oȥ>uv%Eq 8֚%?6PMhe]ےiC@X@Yj1!p/xؙvaw6 s7D?>MBj<&EZe|^}&IwX؅,&A&E0&6Z&%@ChG.RdWChxGEz+5(WfJdi,]'v |$}Ai IFs(lQ]:u%nߙϤRưZ=o~G:Y/o^u>\Lp`d\Km^uwXL^ϴaC u:oSS4V˔jկ\gt6nr͇>?ЅB*~Y|#)M,:{+|BzppDͥX`thKZci.URUʵhCs\Jmd*~ےW3ibkT)G@hҧ%ߣUlmG{p9Iz ZRo#(F(+I5d('t^ !>z8}O@9l5n{瓶;Qյ/=$}*li[(9RZ J:,>!3J4[]D8"T]^q 00d B5 t0X!I'kmgLl0niITNJ Iuғ4'01"JL T (Cz3qv$R@fsIٓO9P};x%e'rQ 39oVMg<ɣg.IRjKQsʘiPGT6c# zTg@\#J-+C)0PKXuf6Xd615Ntc(5HljDMDR0d#&Y33aDl z~S(!!E#*Ȳ+i<\xSfxtΛ~><0Sce3U9"nWB霼l)f6'@`; ']x]pW -wj/)ܟn-w} BH;{1|0SݧTn)ETLU\4vS]#GYL(|u$͡(QZ)gWGS^KrP(}s^u_`rTN({Dp;PyօpҢ}+R p KJ ڼ]KUnZ/+~ >d@;|`"y$'%-r˒VSbկz./?IΞPaMWg?~i0͆&z;  y`ϭsFjZ+|Tr_wZ!̩JJs^M5m"%*B]̹+ւ:uuET^]AueAg`磮*Ν"jwp_*R +pN lF]\>f' qӫ鈁Fڳ9%j 秮*zuè+#;Y9ճp瑫G=Rj7PWWW^hH](٨\w.J4]N]]U*TWRZ9h|w=Bo>Q ?ϑnH~wVkYQ~﷈`Nr4]x nr;,|\/<"zXHz`m몣1? 7׷(?>)C eɒ%..fq~r-/{F绫Z$6b?dI3Y+2Ֆ-Qv|\i/yaҝ FU.6-֋vGNKzլD Io+ِ~Uvt2{%i0o7a\|{$/qx?[ko7Hl#b0{Y۸,r0\:YnV)͌Xy5ضv|zr6w床ukMbWٳ\6捻 l&9PA ъu ^GELqqu<}=ѴK̬Y=Oi_gYĻ=F㈅U.g^mg2&`ѕX 朑|bB!?謓ؙLJr*{v([7p}V[n5,(͍o) ~14vE] \JY3@a#2YY0uJȾ#a q/F/ zu0' j􆐮˖l>N!pOe#QUׄw, Q!$te.(*d ~D&<L42:K( ٽB!P70}~]Tնu+wNXݷׁk7eu jEVأ###tgfX F84J3fHԚc }4÷fh᝹ۻjIÑKV{0xXA2^`"ރӃJ^Ńk}=C"RS俉zԁ+2M^>c1{BvWWGE~cEP+ Lbt< E`鯴rUn 9O{pZ<8iZi냳)IviN}_]'I^‘Ǎ`SCev}:9Oc҉Cĕ<8!Uc˨K< y|? T|]0>D%NN)'OXRU9pqcJ1 MB:w~4^~u2xp5(#76<O];:d ,Y뻫Qcp1U>o<-\M_:#Ξ\.v#A5I*x' .AʴK!LTT}iSΡU%v\>Ϲ?J 2hh*w4̯dRn@ ߦ>fget3V՞LPA{ }#naȶ.,FU=[&}0؎^}Wc :[R趱;镽+|]7]z9!;D6XdԹTBcciع!<%{B_BFC\\NS8#,R.%iŀBJ.뺵j5GiA]RE%H G&n@z_vFΎ#vJpdN̉x>]WlZTSӍc6f5!IStx+JGk0)"9hC^ \n+<ט0xg8W**ѻ,셵RhJl9OGqNҖvMKhq![|ƗWag&Ee@g RnJIZ%U.RNOl>gC#ǀR!8,<%g /h ZL7 'RjtArk7k-ikwpKmz?J9+'[ݧ5sėW:3[IæmzwDb4h6}zz.l?^oF2Һ^?=|wWFnz>cgEǣ[MWx~z~Ro)_ ݵ8`KIy1yz|WkδCsXq|s(w`ܧ臍MY_tc TlfhicE!e?ˮNeWpr 5{''EGYkcZ>E6"ZFIgM]]}BڳUwky>%5T5[rʜcCh.&29DͲ/=:nah.g-OG1CEb N b-u?yzIH4O6NvgrW)lB%^Ŝawf߀R#eُj\uf|x5%0ٻrQigRJfRV+KNze1zAJWA.IPH$E% 2۞$dYluNPeY b,Y9:=,`5ȂʒͲGzYd::% 9˘⑆30+hO(x$MŦ$Bm `[HNXDT$9H/aJsqvIwbMPjqmFX2>&9ֆmT`$5+ '4yHBZsר0"(+1ܧą`[2x$3d64ǰi1qmtM!"$&9hrG`i!z+$ZZv B&|iҁj,9|ZtL+cC6ȹIA+Lil3b!!)5׉>+NmTvI^|Nڤc00UG 2(3S($%GC$e.qӱ8ŲaV}׺N(*kp#%zI (w=^>#zygے$qMA_bz*p Nߨot'ʼpJhfJϦůxnyKvJh" AK2 T;]991 )B7o:܎wۮ+]NJ{c;wy,G06ni /KY/e UF:4ZKO&rUH W*sYILg4ȌP@gpyμ[Ǽ-bށV0 {ɨ=&B o%l١^E L֓tm~6w~,&!c2IK>W! LZKj*R43J }2?{JEJm0A)OQ%$舥B63X/ 0 )|'.p%FKhXg/J۶wWK6e>_EHp Lj^99j}D\Yju;]w)8FR;4,r{,<`ȳHx 6򔆌;8|))3%^L Zu"K:W J%gU]`k{nr>f"k;x^Wލaw> wx; 6[ ڶ Z\Zvf)69HcA&1ҋBjMN>1! As`) sH$DEoy|7tn :|g l]Π".qrnʅ/兔Le UQ5F*Vy]YC*2's[ԉGts`Rw$Fèc31<1* CRi-|gl>u!rHPF1A5R#S2A ! i& G LɷS) Bl{dzQj}#7\?a}rބb~zvQgZC~]a A92ESRp+K&xc e ;rA2o=>0x$ռRZz1o#OcyoFv;/_˛+)~A U??߿uvK ywg*S-b}˃${A֕ugz~Z[~8z:4YbqN·7^OKZt+np Ço5s)W9E<:%f+VQwú}n{ G{yra(J+=]G{cW}}oͯ=Q= [J%r-Aڎ\_ٱˉzsSs?ʚWHzU^wZvD{?HgԳF '?utaRY"R]J_yzWc$Ϩ7qIk>ٙ >%|~tyzuJb8_gõ*璖y֩Zb Zt)G:lSlX-,]+Z>L>t)iy53!"\e[#̵[f?xu:m0pI`4>4k,?Ibfgm,/.Fv;btغcw57Ƴ֗i6l8!t]#!O#Y$R`"NJ|f`Mi1";``arŹv]De8ih'<,+\ tI Nj'0˞跒ޡƵ1:t\!Өs17h\o;RD<~c#GD'L uC`o;|_bȻd=Edq{0[??tx{W$PէeZH}c[ėl)\t/ fu\$|ܛЬ-EF-ߟ]禜,ݙlTz'gEǏ c5̹'ؤ7ۥ-7'/ekW)ˏ?E/m eO=!Ts8O.ߦWD5@?H TZ%^I&}2ĊrRG?{ YnYL?x+zlh ר mD<䖥 4;@X4*oO< WEnә (T"K,~wSclCuTr8v$Z4',=tgwbu4c[W)F?Nv /1>}s9antK-;u񌖪YƢtlcf, !J+k*eYٙ]ˮeWK[v1 *l.vZ+EZ33W\ eB\8!4*k|tfv]A/fL۔+6oj.<)+^2i)xk\T>0%liG>(&W.pC trk< 'OEoˁl#I<7%+:j:`7Yﴳt<^W~u5xۧ}5~EY>gqrRܟhdAHE.OXM1UJiQVQ?7_:4yZeD%a匒UVx^u{ vn(ݻEt.-v϶qeXAh [ 'Z#c.1y)P3Ql5:D֪ +mRV^gsWQG6ߞ4+:za|}9KW’~]5 rsۛW?}7 >_+ wݘg@kk檇y]fglPoya-&Ԓq6}Q+;iP~3ɤ2@̤ښS*e%0LJ&U+͙02mj0&^ԒI#r`eQ|a.u3c/Qۿ%Ųjk l: ->9ye1J:iAO}'J{{ {^o:P6isՈN;{ҹd<}|N`&senx.ZǞ9[3e~+@8;vq6kG 3P|JcKBwy"dt:.K,&ỨjҘPEy[\|7s ٠+mqb2CYKa99״>qk^ڂ}]w~s=s֧߯Z0@f=:?c4EX'˴ejTfE#zwՄwKd6!/S;BdeX;^ A1$|flI$' Z|gAkD)eJ%qZ@d D$'hPc(lsJI> փ=_\:#gVva&%[Gŷ"gVĺK37'6 wJ}36:9;KO_O]O϶"2fs'5DsuW xSHK KM }SY&a唶VR-5|X^#Ά޲zY ƲҌ=sVA&IH" C.&E FRDY)8xѵ!)l78\7PYkuZE$] IJNٺ 1 5E̫Lh ^qQy\պ^Ie@ ]wL6i1jDlVk-%3dT|$c#jTX5?e;m6BCVnqMm@{En#sڻ|}"Zx!bhP()D;یtgtiw}fgNiZ ??{SܗHHB:Ԏa-jELe2#?8Ѿ MFGO`PgpAPGQtJ: tQ=ٜCΆ'vE'}QBb1KUE224V4{jFΞ%6e~iY矷{gwǝ-om =&yq?t؝~Ϊ?SV>;QwDn%mQ=?bǤcNO#;٭ui=ޯf{UB3o|7ϟ~OwLw' xweK5׷@enyLV}E*@>F%:7,6"\K\8yf_tQ_]50ŌUǪl0hK;兙v@eøMvcDK{%TfWkM$>lBB9 uΰ`Pޚ4M%i;5<t~zあ uAW"9Q9+WtNF Y#,r43%2 SM%uq#ɔ148DkK95;Z/A>w,4~>3-ۗ_+bjsyShUEie 89"Ldu?fG5Q쵇 y OUCUwS;FH h8^?1r<{JC. P 9&\3-+#P :NJtJ$ͤ~ Yq$B mb@f?Z(UEԃ֖iMD6җZxՠ:/AH*Jr$"H`^!$%(E[G UD 9)_@RʱEjBƳo\*Q3VKZe\囓]1,؇zyJ╒~GOc)y5h,ެǟ89F !~8YuIiN}zƕs?7Tt||uir3p^?-8vj80[e1x;=?7: A+/ޞ?I5G oO~\>h*j(͘8=2m~'"9Oq;s%jXoxN5f~>9WW4P7Ͽ.{ }}oq;o~;kkysz&v#6Yywn/ǬYᄆu#k|&lv>off :0's}u>wX}'@>Hr?'YRbd '#[mYP#Oׯ1^\0Vk(T[U[o#xhF-uW^k2 REF Wt-ˀ5ѩZkƟrδnhZ$IGQzy2VgoyHμ̋z0k}a~ר׷Lo>ӻ’]rz"/=*:{{]GŨl50zCja+98}DF|zh^ CorhrwT7vk ~}Li+YťDix%'U;B'prmj-ޫckX,xAŠ#PGܻ,`dVNM{Gۣw-<j9] <2IW+h{{~»%zVV',|޳zN[˶G=ڈ+CEo!vڟŤreE!Y'ٲA bK)խɗdKt1Ob?@4gܝ|v}62"FGU5:Q;kC Ri UL];Gӛo]LX[pX5,4c(uHjE?MDR08N {;l9v0v0I-T+"J Dڇ䳆jQGUD͋z?Fz_ o/V:c ^;(Jy`KTXp,!KL_ sVP[lԚ9'Y"RZ" ߑý~dp;u2ٮOd6]=+ ?\].Xت;6wL"ފ"('K--ET1.Wx(12eg\3Gm]7s?`kjA6Zmnwӈ]&b4AڒC +cWQ$^hB(1Z4;8rw[fE8{"˽5~`ۊ;wOUσFG֌4L8EPUw{OvԘk=Ж)̑0N-c:Yertݫ>%^Vxvn`y[c\LjntzguLc>3\$~t17k3j~WM[C<"jj(YJj|W[`Hx4:t+WH&=#j]Wms88@HI|j բrԾOM誦#0T{VOSC@Mڧ[UDO-[]w9|!8Cδbhi<diGﺺF: DTkL[+;Orodk:xa[5= 2XR,]F 8AێDOJv(~ m`y@9o[ixqk4iFz~HIuLߢmk=x%E86Ёh-֪W>ctd)x5,EIT$A s$:d~=E{s3eaH s|XULhoO>Y\p-<ݏֹ]ߋo,&k]R~Uz-?ދ˥.p!<74_]^zwym/y\->_l& ys)w>T`#{hf>k|e~25_dc P3YJǦn?1\{9$W_pP8ƞ[?o}!;V'ö mZ9rsx(7s[mucd;N7|ZIZ>Q6j*6XM|*`|=}K hcLzh[5m}c8,O]mj e*k:WU(y=u}W'/[؄L!Fճٌ0un<Ylľ)H<]AΒFq>aGdyqr3nU9yÜȑys}z(c?]]sڮ-ߦ dCܬ_/5LwZ;/Mb.vDc1nz2[򤟴[5x3پ׼]}8.lN-ʹ<3V~qC*q+B{v^T$vݧoZW^}s̻K}:$n\QYjPh+k]8fj|m#bBTj"hd7K=rt)L96o,99yzT;bJZ~DӤLv5{{6D¦BTPCoQ&>Z䩍 ]Pspv$ΜNegsZR6P.|Z=ɳ'~p %ñ*4 V;(D'DPp"(:"(%Dp!FBMFrlJpr+u>uJ+r5=BnMn.@ˮ\-'7I.Z^ὃuf6NP]^\/.d9.c+ pzn y8v[dȴzL mH>S('(Ns:#@zs+*u(C %&!\ƻbZ6uʍ"Wӑ+PSBȕl+ {WLPjrg c}y%ݒ__YoKhɐWmy?Öx_wU5m*4:Tښ^yzzm+}·{Wcf&efgWhu)BV9wmMO' /ԍF{셺q^WGFQn&!W4BcM5䊁MFr%"WBlrŔB ʕe=t Z[/?/i8b^]UhůE_6R(T݈}I H'ڦzuFuUH6VbCsmeG$[C [6nuӢn|=4J'/sPWqٸku.>]EܘWq:1 9+\1˕P6VY y#+xWK!bZR>uJ($ (Ͱh50'*=jlf[ÌTF.ZRWi,N$Uڡ2\ pg̸\ S?+8Yj:rIlJp^rŴO\ %"W+BoUNr%G벙 ˕P IUBH88Fc.r%!`AUDYTc-@8\7F.+>\6Ȅ0BBǚl02+FȕEJ(+r5Aҫ6ῤXl0A34&hW"h.6Ɯ6Fs&^;r [y Ze ZT(=7onq\1EFf&ޥ.WB Qۍ.Qn\N3`hF4 BF*-G{دQWi8YTz:*m%N[ lU>s`EցO]\9Od|Fr%!Y Gktr%Xjr iɎȕZ\ W˕P.r5A"BrZb`|+gAh=.WBZyg\1p|*"WB?+D|%rJbzDkWc5j4.sgp4;hJ\ ծrcL!xEɕ|Jpr+}B˕P&b^Jq&5]I5' #UIeWMj7s&5J ?ȬUd`O&WQpC6"ӒNUʍbUzi! Qȕ".WB .@* 3e< ?UamRjUdHiJ .\TZh]H]S9EvCʁQ69li(5FSBݝ\yprXg$W볙 mPSUS+m,Q:\ƻZcR+DWjr4=j,09\ Q˳=JxW_\-l!kr5Y:u8Ji+(rXӃO^rUȕlJhɥ.WBB ʕ7V(2&,7AߦIO^""WB뒗+Xjr5\1 \i$? J](Wd9yW lL>r%6Jh}[ LPvD(Wpɕ|ׅ֮\JhbE;^(?~JQc>ٻGkqjKlJ+]걦^䊁vѹȕڐ\ /r5E`\$ڮk9_kr27*32Bie3rl\E%ȴF]EW\ >k}劁#Pr%:85lC * pkhˆvKJg8>s(\griJL.egiI9\108FYC?+*]L->ճȕ#H!(`.\$Xh S+ enr3A2+&i=v8ZR+E(WD 3+&Jp+%H]_jrQ!Q'4Apm6g}J(=jhzs8`:w5 ԑ#-4-re\=ZI N}pGy2+՘\ E(WyPNLQHCDdPpu6߄_J[jr! ;Gf#*>sBIejrBI n>8 BA _ܹdקo>vνT-y9LZJ8Pܴ}pӋON!?ܩjݡ8\%=kQZJP7Mv /]2zxKO}vvuso~Yٻ6W?%Ү[uK`a$:,9,ZIڱ[^8c`lQd[=N3(fkv Jv}d9;-f7[f֩8۝+]r}?qv}3 !gͯNhIz<~;,Yx9b?EzTa|?q+_vTvTT.@$P:|BxXͧ7R~f2o>`]!`9rtyYaM[ź\.sŸ6ƌؒCIhz#45Y:ˑQQ6݅DP2_Ǚ6Iwl}B{~>ߠ?jUZb|UۻR[7{cFW첥aQ 5]:GSf ͨ'R1FP۴֧.nJܽ. sGa*d -OaAqVPs#=3lҩL>V Ɯѓh9R:o- A(V1K&ׂɈHc5NWJ=hR$Z6ZD.7oWTSԻfmV)UwPRyTke{ Tg1]kGfhnE^3.KNI9_ᵯ&$3G_>V'L#MZ)ʕ<@{J) tEv(41U`!w!0XU֤{{1;[_H4R}tD]FlTHo붽HU!Gd3<%cNօ9 AO/Ss.ޟ7'!XUj5ԺsDAu@r=RNZQ ~EkjaN>,lGtSt-!wf8mI!akXko&FVh\.`5 ja}ȐX8pj6z.WP< Yb1 UW)6QO )ܔ 3HT'*H`ȒQB4BFh եf ӘJ2P4@ZE36XvseQ[UCEEG |SZXø:hӦG2~4P8jeWG(qX B 5!2ѕ< q'YgeZjh 0ճ55i0o\WMU֣(͛հ֞_5TuAPz*'*Fe@pI2)'[FtTb0VjXq2 NsA?qN +Qs (MQn2ƿJt*TS!z@eΦ$X22yiB AvEc kze ++eq2FN[@h iNfePDfBEA31ۢU:B:oY{$$XYu!LX;*q b`.Lƾ8 :)$ԙ.QM8uT&;9F*AC5gJAQā.QQ&A*x5<;"KQвAw/u4U y +QWaĸgU]%٧VKx{F3Xy8׍!!_Μ7sLJDA e`d."2Dbѷ{2{xWе?V|p.#iҴA3iPc3RU@1 Uˉ8!bEw@zmWqYs.χ?GUЍ+[ˮtڦJ#-A7 /Ml#Td &riU=$B+ut2: Ρ&i֣Y A9AJP$rAVW2pȚ[CŒJ%$0A#|Y'H@ɘuk< o 7tXtu~V% 9ՠQjK-"Wj=xiMR}Y+$QȝL;d?,:YJև\(T:cKе 2"hѨwPSvz d<Ŭ}(Tʗw}CpGM} $$LES{X4kD:ZΦ$vH ԁ>:AZx _3X-D;f broٙ%3[$ڢ"`.Z0hMEkaN3^F\/vO34%0hGĉ4k5SzB'a9⊙d E4XoCn֜M5bUri"r9\PY5iu.oBG!&MSZ,3wQ@jIX-* mOH]A#Sܱ ՠ?lXyj f|Y2ˌ[=W=7/_|wQ R%D]fKCi-_~gGDZ7XKź+havIcGZ~^^yym`B?l9כ-jܜxA?WCg8Ƕ;Rq޶C﬊g_m= ϯ}߶c܍u]0~j L1\*,53 WE NQlm<'z>'r $NߍB$ $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@zN B'U|D~{'N/[; mqKP$/p83Z@VN ~kt}99'? 4'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@O 4N {:N Hv'ZoMWq=E's(qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8t@˧@@{ @O 4.9qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8裭PݟZ_^V1v{saCǿKkuw(+L q9%ִRc\\R'c\ZqizS0.j)2}t5[2t5p*tZNW% ]=Ed-#b̻Qi_lҪ]^"_!>Gxco ~CSt#18Tyi2  ,V+e8W^n0q-1vr{uy1.4*qOtJFy*N?,6//Aҡ/9dty̘Ph{f{㱯pو;z\]`} _3 RYc|q'9VoYmzFe%lӧH8ē=8GH#9TdEƚ+|:ozju@S]ݺ7aC/ͼC*/ng[uvt#O^j^(ל)/WѼ>cR7F;!n<h_2PizgK;%`i\O\jZ ҕW)]`ՀKTj5@i ~tP'DW ] |2t5кGa >Ed'DWŶj%w*t5КGjd'IWN VPW?ǕW# 4$ ZޮpEMKַƃWe7.6?>[_!Ggi{K\*ےUXJ:Winۏ7{ ܸZ^o98Gq\_ _uLܾ7nT@__?{Ƒ ap .#AniqM IIW=C)1gqYGOuףU]5>^ak{cfyujʐTIt]hȕʠ,.R߅R$jn򧿾84V\fI)V6@G|/Wzׯ rŒ z9ȷQ2vT4mfnڶlu\մg`w~a^볇#ܡgG%kXzLԬE V\%rnJ*tsTj;sIgг_i>yMܫc:*qt=Е݆lYrN:lM2J͘",2;˜< Ug.Ws'pSCDE0/Xk`^"DߎyJE:a^|DKKs=q!{Q+oJT.#Z"snJ\JT>xsUK4WkgX}Zbb)DL'5jmZPKp:DwfiA%SE =K\bZIĝzJ2$mJ8@BAWZxsT1Wzj-2W`-Zc\y[UᦛD%K4WjY{U"WDRM7W@e\\Y/.DŽ]GVڋ˜-,Oe(ysJr`Oeay%yzl =0v#+pDe3sW{435*,QB:Dd+UEPF2b1kQܥ6Ӑpw9v7@(E!rE؄ A2 ǨU1kQRk &z,@G*1ޞ$/^#SUkZ8Y^14-l ̶4q>;]~3v4|ya_ge\F4oGtU:]xcfWe;od2\`䋙rZre$aTTnOgml>*~?o1Z @(:v:Gi$T ­Rii`d 9a3ࣗJMBB1č^ywۨD R9*H,NY,ZaaTLiY4=@"SJƒGQ,Pt]|Nl26lߴ \Ɵߔ`M.P.0.){p;<IX1!KY .RkBǒ0ΙAɀ :߽2p$ `ّ X6i,:2O@§$>T ӫǭUUb:Uaͼ_ssbt?;OkmоJ{Wed 7Sx>yaA덝L/uf8L`㏽s8WZ>f[Ye'h{" ]f7ӄ@k%)j8h `ݠ'tzl,vzKгysp={:U0>~O2~ŲOpqK< 0+ A)GbRD9E81SyT0hJ"ΜRӻϫ<7׋kV&M't_dJT?}o0ކ[k嬛%sX~kc_r%gRFj6N1FkEx^i(zWh֔/;N"N/19e_#>9SRE%%":f0 ,uήR_e45 pFKSvK?nnq񏰋0ڄje(sw,ɗ&K?̇ӯϬJG\$I>@v$C`~,I$XAЯaEO ߿]\V/eZ>xvyc߇cgi!y_iŐp<}܇iK\=Wkw`{#bˬM3ö2 UO]O\`9ݵ/&>L{+meYLZwo[$Y =ꆮj^(9+DpWa*m!HSkF-6q;ǽ5ɀx!+s0ݹNBkϗ&ujDSSBh :mLE0V*R&FUA[-ղt6 ۛd/&M^sb;fȓZv:ImU]vԹ-:V愋D.fm9"QKTOHTR՝pO?@ړ9s\qes7DXCv'bRcg6(Y8StTn=4GfD2. +#1A&©AEF9oK׻A9`k ;ĬW逰 ߜڃw < 5XSe5u[n(hfg(ߏNJ tPtpQ~3vbĶ8iip0NWY6S >~M?qn ;Xπw{ v{{rr윹39si!P nfI|¸φ#~U$ d3k,PV8dxDD=$~Y ֞b`%pZ@p(f$Jo8 7Ay,r (Pʍ7 Qf˛nNn}D*Z=cJsB\rxTb(7q"7#ƍ^-Zp=J4E"h@ 8~et F债`p'j[tFg}ndtIAEP(C ¥W@*UdZˎДP}>[j=9V "}$V(<\21ʱw:%b@*M"P#(T,W--e Lb +"&^ywBDD(eAZK̻sL`i'c-I9yҲMFZ)Zm>v G1Ίo&vIL"E.Ypkdw*r^gIŽ]gůN>yu}#6]ɲm0XŶӴKӒ\/n{יNQs{oV[εʧtWaHqpN`Yeo2UM%XFTpD9 Rdh07|1ۯdߛ8P>*m + \PH{w-tU"n*#WVN5$%! fc`8T3a hHA$37Qbǚh18No3HY"!(R@mrHEV*dgneO|rUE߰?8͉r?v|lL 9چ$"q҆,~-2;I>\/;Wgen%GZۛȭMsc[mVw{gH)붛K`[ϭq;:]:;O[زun=[=/lyv~ƛ~>>#{xvAg Za Bd󖢊#s\Fˤz469$*% fc; Ep`ŒM/)tL?{s$-@^Ps O`qIb12DRJI΂AJr_/No% EҀ͟>rBZ9E -Vt&|nt,Y |ŕsZ:.}6lKԂSv~^rҖPl+0yr\Z-rI5rA"*#n:$ 'Td 6*ԥR*PBj#g tư18 o Xxyv5 e{MW ;~>E#1u%FDgBuMhU N1XiCjvhTQ'"6hHC6l`eDtv3 3@IȍL>&n#g7b0mAƸPV P`3$EEBi0 ܔCEM~4pJ \$iA9ZȨRVkB$^24 =RhQZ`]H"FuT|K~^6c[1h ";D!pK2B+)UP@ʉ <+ 11֡@4{,a4ERG8 $(bP*yQ"Q{ȸ,>lHSQ-SWsG6u6%"@ø;\pqkj)-4dL6689Eh&!LOhtx \l 6-ú; h8R^ۜ:gGr6ZS*\h .sZЋ\Iz(@F]pbet`AS46t]ݧSK>i1SĚ>ujݺ3"tq H姬*n 395oX:G+8*`qqst9\}_K+MJY K5+w0 HWJ)/ ,sϝDgI! cP8 :8O N 1" S4h7ek#s]QcU-6pcwNSln{~YWIW?6^-kQ>;b*r!B re&Q͓Y` w1HB+ʥ5Jk6n0Cd 24l8xgU^ب7_1ZWWqxȋ?o-a(o]D2,{jǿV߯UQ؜#ou( ާ:Ēҟ?*jW%UC 7գ9UJ16b?}QH]yjrNy/? n>@l\5rozb|Kznz;7h6(_Cz]7ZWkMv_R8wnw}cؒ.Ǯ~T2!+vD"D,z"R]uԫYo]ۋa^J\2l@l] j/ JOB27OvEv8~;_{⾫^Tf\uFhO$,1ZRO.9I4%q:xJEe62v*Z{ѱ(;btŚkRQn5 VER9daDgajpXhF=jVzݍnn8PXۣ .UYr?#MR0'yl;0b1o\zj?{OzP{v>c |Lu OŨjCEzTeU¯d.K$bSA7MS]Yڅ1O|u}CR~^TbJj(+9e˅5?Ԙ 4d ަM*lqeUy-<CM 3EӑI7Weܮnjg:gYxwNG&3OgU}q ΰMOx8ؤY7Ij .;nlf"أ݃܈KMY7!vڟIȢ#|k4XK4%E"-wRۺeKoHt2O|?rhOKaHn>H3\fCkK${`Pf@/sOr5ΛCiV/l=V9ݞ↡^9 r^*^U]3 7?U2+m/_{W2v2D\^y𫗊RP'W`N\FO2WJn:zp++U D \er@l2퇫L%!\ip:!WiN2 \ejt~vhO2ʜ \!SL-@*SYWOԞ^۰wu']o WwV<1x7*eJTWz8!}2pJO2@WJ:zp(P4/fA.|7pQ8,y1 gV*0̞8Mg}tR` /7uf ~5! 8JݲCʔW ĿF_EoXqpוyS Q"/'PtO$=3BUgRN|wI_.vl&f+#${2 jRփi3Al6I5GKtq*e+trvJ(N ս4}(gΗUeCl!們! gW f,708g0ع@1deեLqXt ޟ=aU6qI Ƙcb6#^D園>g!Nu*=lW 8m ⓘ0`dg*%VUBhOW/шUtg bԙuUBI{ztŤ$i'0 2NX2}. eTi{'Uvs oC4 p1UВ; %w_"M &w=03t JhUʄR1HW*B'v\ͻBW!$HWa-HJ1&3UB[o]%^"]tQVD3t2Jhۿ"PJCWsĔ&o]X=ْygښpʶ}JAWSc$+C?85˯I 5A_"<ʡ7_CG.zS^z{6qim&hFe Yi۳W?OF]!“_|Tth/bjYs@fWe_ޞ(&d %4BJ' .M|}o840.IIz7au?~//u^:QDQ #O1Zp5Z$I 9l9Ĵ›շ\s-j_*tFiRo %֖;bRZBz {+=Ū}X ^6NpgI$5`bP"eG 0wJ7$OoC6=mf +3%Vg*Uma=f*eМu*ͯ-:5cQ,* Ɩ਎EU6cjҩݷ2b4j0MR{$VHWhRw4wu4lZsc ܌,1|+ĩ9 >0>n7.m֑haaS}s&,Y/2g~=<;d xnEҾ G7xW 63x +u7 S*RѭS݄ƻC]GNI*H rD&:̱k˙XLeh]= YLp%ň}2l7L'Y$2![+#*hA@Dբ|?" 0Q0wH,* VrL>lBGQ`d:#6 h$z=%8T;)rZ!۳`[`׷gͤSgx\8Z\^|7>cZkֽj Y]NIF6R7 ~娔 HƜDgC$B_='|ʍyDU[W0O,W߽lYŎYiYՋޞ[0&E] -C@eS)(BZ2 Sߤe͛fgyb26${>YݺmOܱ*ِwߛ~]d% Kȳu18ue kJ{4T|@<(F3pW ncLeP屐hCQyWau@S؄W$.d+ e Ql%#2NYU4`UDj tKJ$(p$x@Tvmvq]fׁ<3`/àh*[F>Lŧ5 T&ʔjYU1].y2DKG4+E]*CדbF[pVAZ?:ɖ\-RՏ)K-hW&]#M&P0vYGjd~sVk/;0G[,} *efmOy2~k/Ka*gh2_͝uྐྵ uTf a0/C2MaF)$e n? \ <{w5<0?N_1[lbŬ,Vە])|ǰhF[家G7fv&2 ~~q77HwhՂ&]^a2goÄ>݄ToߍƮye |U̷rv[[&ڷٱ\rڡNpwIpigR;Z&۞ڙPؿ>iR;Dc\M(bxD#Ӆ1ns Cmt9)$,9',"0if``M.̎&XF~ڿpdWkTҷ74_,=Qԙ4{,\|(^٥{A6݇PdSUZ?9\L. JgPh*W9t]fwU).1YT%3(-~^͛‡+_:+5j/i{Vw/m\GڮZGTNPZ̈́z Z4W_RZ(._z?e[awDUϭ%>S(QIZ9q|sKь ]3K77 h !cT[YK3G |y$ri]vS$,ͻsMeL [D0D8,cR!( ؂-]]"O~)-XvvMFW'*P&ԴE QiIBpPB *G$x H Ix Ѐ]QQpfqL9UaaZD59}Zf`ݸsgŶ'[4h[v^.մȕّڲ>Җasa~h9+᥉9kgm(9n)Ɩpb8M0D<ƈFQ(֢tPPG POp).}Ԗx '8L)( \0G:Hc©Org#10gĽ=3 !EtcIvSo*@H%4 k6e 0@ |: )&` 9P@$M*w1ffď/WKmu6&%'" yŽG,)asZ'qa4Z+qmanP@D3AAscb[1h .;L2/ ~|Gn}G?_ߤ㉦Hu`BFhFEPCFd q,m|Hߝ=XI+Rb@_W+VsӊiF%RQQr"`!6fG EέRȰ, Ƹ#aREGo6Wsvn1t(ӏ`cczce9 ߓqy?g$>kdg+o`JZ e ZJ(H^imH?%w@\&˥6w_nr)k-ZRr\_c8D#QH[T%Dn#! ʵ*sO $}dTQ^ $L}n`zlG4 >6g量P~ג>NC R7`xvwsZx-R;Bo\U~{y*e@B[Rj%ImM9r2Hv7t˞#<Gܡî{1T݂R=ɯPȾ(k]RpMߨhЍSݣu9||LK(1t+#[F.O7݆A~/0mb sLg;RF٪w:bzoJOLjp0s<:j- 5G?恟tg7lrwxK]elmgd<&mG݌U/ФTq^;);zZuC87Λ]7-%g9+>hvq_L*ASv(x/YBlٸw ׁaK):4xNA| //hhnwI;e.J S>e6HEud| )< ӫQ3g;nAðzy՘S) -?i,Dv,ݹgul>{7}j 9HH"u-&8/4:&ޘP1!1!ZOIz-֦ޢeQlbJUL&pܧ>D 2օgA~/{(-5Z&SCa-p}:OE&jKDdE.ym[iu_1?#Kyvvv}"v]T^?Kw}:`*Fckh2 ,! B [7TOzLSl޹fR\32>_fںν!BS#Bˑ;ٸ׋=U#vYGg& kQd)c (>XZ/) Ǜ9g#Ǡ71j`7){LhM.|u= ^41q:kgIт~Eb1勚cg?Q[0}׆=2fo\+#?~UawT˫-\9E3#wt l*/o:ŕhUV~ʾ?g惁%I& 톮c?!囥ϟ4 7t'][헭Ĝ,&?\&} ,thK9|ӏbMj+ m+`o'Ӌjؾgh|_ϋ,-iӄDQjo)l冎>Ї|-1~70Taafvén^CNh2YKXB$ J^8IAd}@A2p[_aB$JTJ$ЖNC u)`6)@(uAn_9vtvh_u'DZavbY~|{%}(-&sq7V]3@Qh`HJhev1C4dTo]B_佦A zZhD/̩$K;|%%ѦP L mB<{bȚג4F;:.b$ X1dy`%' ۣsj|t5tu|D3|&BH%{oDJI#Z)dV\v dViDE6 :տտooڜ8'Lj AvJjbLjZcZl6(C1Dz)Ur8hLѦ%\ 5 Mv˳p`*CG6l%͔# @;ʻR0@$:c7iXZh)md2a'RVeuՈ~hkܮ{U:&sHb|9.SVЈ [K rDU(͑@;8%H|!S"$$;EJ:Z֓n?)wt vh<*9ES P90i06$#)>jV$#a[زUeYA+Tъ<[UA҂  a`:l;[YKWd,tBc~UEj#F2^yNsQҨ\ѱt!+=HRCgODŽ1-NED$NE@,rTS[ , tILѠh^NՊ'cBIOfKh HҠ`*W"☀jR!`Mf{1UpblYPo׀*+AJԐM"HЪf*E5i% 8FTm= ]rSKc0c!PdGbX[{mAVj&:Hd$BU%f =y G+eӳ Cv50uWAIӉxHY5xD @Y$M&DKP>L?;^:Xs'M Ff(!e1X/>.G^EJ.2)k1O`E('sIIb%fsy;\׊YemFBnȊdUR IF[uhM"ZZnT3M* >[hE1/~UkbtDyWxqe-PZUG&zbDrca)"{&l+r{5QlH5|4ĺ *p,ֆ`. &/Tb8@`mE`ߣ>rtDd*^@(0~ZT,* xə@s$'a i+Q%HJfvju*K61 30w &GQjEB[SfBܺe%2Dkդ&Y6n!bK_,$gJE-PcyGRwY-}חkuܴW jwylN<(%!௿zFkӻ/@?EF|im1Y@dݿ4/M ګ~`mOj14/T?[l՛usf,ްdgz(\Ǿr!y>)?6Ȅ!FKҧ?:b\fP$-&iaq3&ZF&>O&]K8!Q7Uߝ\}qT[' ̷qAoǻuH®[[}Mo|ΦǬ^qZ+rv#BI4.-!غrpP+PW4rіOAKw !6,$MNӳۻ:HryNw^~yMX^&7&(%M/Xc ]Z'kjڵaVa['_/rKkݢo4CשNz؎晛Z R"4hehcg\rB :}lV#LQ9]c \΢0@d&`ɕ@α%ּLl !:Ut`8Dre|AGG=Mt(Rr~%P Rh庋 y \ Ǭui'j؄L F&7)uEvнdڳ !yÚ˖e@Lj8s4QcmaKg")!. eV[ % R,ٖ -k)>PXOl; Z2Tج]y`+UX>vY]R2 ]u+iGY{x켢ϼ'|.S|3J,zi4ڋT# EX8*ю*GR.(*Q Y Qal `GȎ'DAU"@ }VĒ(bW4E5F% wm&wjVC[ds+^Kr1hCR߷(H"-BdkEi̙iQ4J^%QkpDY#I$Vix<9]|[~6ݦ/kZ:N> ڼvX0{+}jK!;V}DkV.{|@j?8~ƔGHLI4YKy݊Qq %e3zPrJ'tj+B'NuL5VQ-\ oԣnnDqYdij\a]Ib`Lhb"@ʍ W۶˧`)ߑ o)85)ɂrTAKs᎚$)m9o ΎM4bXj^{=J2!:ΆW+ͮ}zkDWkVksG*E[㙋SNm'FL&)0 S hdiiSJ4Mk7+TITY?rgeia2@5c2S:͋},WkpuOŦ\ےx;mXmLlST{߈+ #mʨ9o09#furزa/C!D&l:j#AZRq'ylqf KL$/[JTGO"b5<2\\DjQ\ Aǽonttc+zx;M7u=J!"EX6 _}ph?wUz?~rs6/}~yR~lIlG웪/'- x65fkx'eܷ+,2jW(ʵI%RI#I*kUTwrj_i RxX9McCKxrĊADwR˵?n=&*d*d:4O8NDYBq&5ɱ`mD(\hWXY _ RA@ؤT gIvbYkpNWEO6-6΀jw*&X?x[eso|[љfxs'D;^`tNq2,Q17iaP eNw YYSՠlU_:2xL;*NAvqI%` ̈y9r7MUJ|)^l(2!m|V-:pȄƎ"q*SL4,t3:=nդ$`H!* Jm.rs(@<7bǭ5y>vW.o_|0u@(l']_G&yߗckVNI1TR"]l-1/fgo$f碍rhIs\XIGcADۂg1HE/]'`NN0'޵ {;T9"(MU.5<&eu"$1R$Ka9#$MjڐHͩ1&9ͨqV$[Ms1*+^'Y HH]=NTmg~:˺G_:\[Hr][]u#U%AT֍ZJY7{Y7g`we'S֭1DWب` Z ]ZKt(9 ,Z]`#94* ]!ZK|+@i he 5+5!wB]0xte5 0$BRBWVK Q; (Zt !t-G((ٝvtS#RRCWWP Z;]Jف]=]1ʅ].Qu?͓I_׹)|Mꖍ(_rڬ03z.on5_)' GIt?Sk|3V95BGp?yS!^^q.'AQFKbX"\Ld hա"'(9]?1+>fx\Np0j`,orQUU j.3mO]x.]> u Q[Ug wB.5Ẁ^cMѹ[/46eLѷTf.zkK޾*hl;j8g Z4ʁeқٸq# C TP)Z+Kգ{+kuխwE\WWoU1ql'XqTPQG_ւ/o(T1XKW8x7uDb}9^Qmsӛ*yx`X-±[P}R&}a"[ & Z|z]sAFO}keLY 8eT-mbڰXK#\BaiD,(u5u,a]`Cu0tpy0thRɎN4 +l 4Е&;]JMXGW'HWRjI@t 'DᤴVޫ+DɻS+K`!e` Nm*ueUJe ; h+qFAeGfhőf(_gsQڀhGWOu=ezP0#:BBWV0 Q* C7/G0ֹcLquMUUzBF[_s˲Ƽ}`]]^umuh٘pbaT1tVL>vxP΃)ޙzkT m sX$V6ɋ[&E1!$: F\Nl(22RD)X'NQ1kSUE_"rM/Zh%aEW$j/Kw٪-Mٰq|x2nδ්f5 f!,whthVp D]`.QWP j;]!J#::Eޖû-"jL8%5ongO^`;M#\Bi@k;dM"M+ IU*n5ewB&]]iA)KN }]!ZI}+D=S+#ش`ٳCh%%] ]YDCWP rEttbz'r``yfpՑfh=5Ci=+` utTSɄCWWP  QZ Vo1Jv` *4B[ɦ;'ԂGIɉH*"72nWP"5w( JE6EWP Z*|+DY ::H|-MX؀}jDVnDXp~/AǮ kйI4Ծet4,4-DW ]!\Lh~=ERj]!`]!\ ]ZC%S+F gHNnҡ5wBttute)r#` `r}{WҲN T 4* ]!Z|+lN;z9twt=È:jc7En9ZySX ހxGWOu=5RP]!`K++ -]"]aőV3 ߱2F׌Jh|ȝvzN`5`c)*E6r*By3ey'(JDWXpR+DkJS+!A*(i-WD0-7hi`ڀhKi{S㚡=V&I3w4}42DWJ ]!\ORJMH"]iJ$I]!`˃+ -~Qr ҕ5eNWvtute)U!XCWw^ ]]/1AFG1oV#4C)=SW]z]!`i+kl(tpb}+D >]= ]1ytT43LP:ψ3+?Lzw4KBlӀe4hiX`"=+׷JIœ\Rd5ЕQ Pr >(B3P>sYN߼EK-L%FN|;w W,U2MLŠͧއ= va"ا]Btf 2NN/G w.r'TݔY/wEr9ZyAE&~gsзaU0c~'u5IۂYY{I?wXطU} "%wӏv]\UOp5oKp>b\XbP#lI&͇_)o77v0װ`,Yg<}U[*SJzwKBFK߈Ao3|prF|TMGxQ4Y\{>!j-ׅU_NsXELCrӗS͝mX''<+ؽC[0=rvϷKy]Y.UC[C[Cn-3)n,2<$eS}fvvoRaco+ Gkb6JHr^{$>}3c8 \F4b_a 탉WH!lM v$q0id륵3xS1>" '+k?nlE'Uf0WZȃoSCzۙlU\U6|Vn혏n*kZGwn~/q'Skޞ:qhblshCڈx-Heu0֘f 9a2)3Kby..c",1חtLH0km+*pG fLӭ?{ 3{})|qG%Ocd-Inq-ܗ!Iӝ~ql%2%Yv{IlS,Vk3O'r:<a9 ve#kTBU\L#}yF0ˆϛ_" BRIâ}X@+/Xcќp1b5c^T%ޔ/п%lXf 򓓑0keFnɹys:& &?eRv(L9{/7_ߟ<ͳ0!.3XNh, ]t0ıd k E @یQ'Zn*12waPWH8XKQVYƨg&[ϭȘs3֊ȭ1/Nen"{3CNͰHN?ONGzERR,D"lcFnɁbS å/c}Rmf-B ?9 @ =s:6 1aFֱ7V(XEӋȒ[%P\rk1yor鏦=dpwaNW>\Zs]M2yzHe/..2+\?gDfpC\h5@ sA2Edy{?r]R'k[@Ԋ->xVLX,.,b5@O #2*19\Lg傂W Ղ9V8QPz{D`3FQK╹S몭wc_ 8*Rw˪|U0q4%tV)H626 t o#,J#"E*%Ԋ.NR2k11 *#$In2918`l7ԇM Q|qH \RYۡ NI>AHQw 9f%0:(tJQ<"ݶʊ{ " J~uv0 g@swQrZ!+ywqE4v,:6>1?Ǩ۵ecLq秕#Y)d^WTPgF̧ b9sh|-T/$o0r C[2jBeAgJB -wQJ;gYOjjhEUiI: FJ >cZ);|f"Q=fU; oys@g˔ʜ(#s5CQᯓާ{D2&ujEZL 10R"88Ў,\XZmVyu&S2 E1zL-a ]Ǩ̼2hkb ` y6CebוߜF Z@D)~~/$4(/vfLX!<D6꧴v3^YJG:)cuq%"7P^z@1;]hڶi0 w~9 1YRלZgmE >DٗpABGB8BhZz+Żu"\}j#| -2vA`gqyh+x"-[q+(56+()mTpyv[uLjua2\ñ:ǚ;KuGhGO.(=9nَ!ܽ{0q ڍE iyB Ti!fU_v:O!߲|`2VG|׶lIѪn&y{Dhjc3@R\ GmLt=rZ`^Nn(;؟WUٳ}ww\7OKOOʧ_oyPcܢsakp$o?h!% Hxd 20Ŋxn$u7WGYDs墎D[2z8Eh(8-Ф] ů%ha%D@&rA3+WޤBσ>mcWPZ=á}ws䮧0P%U=ň<<~ŷxyp.jm!NmDh3PϠEVpC=L|An/C`u倪Im*Rs^;NQ9.d)kb@q63}QS ю#@JըeԵX.mȁ9@(0p|;"&꾅doXՀgoPEtb%0l,R­ySGrH#c:9F9x)//FBֱncIGl\Ę2% ,[ոoQX!UNNJM:8 qi:ymO &IFByh2^#h[ni ԐكMJ@/07i$#-uڤx:^Բ.$@\K]*k"֯yCiB$e~"kV4NZ&Co<'zʃ׭㬸Yv7g+k $09 WdDžax 21ZwC+@>rĔ/f-}&}4.MHM!Lb2ڱTG*eb>y}䔚oITJXHjfLR06Xe%,W2bfKKٱգy ?"*+!&P rp=5XCoOyޙ— TC0͵OB}LB~sNG;d!TQ(ue-/K!PB!~K.J+$9L+#\1BBZ O۴K6A"*&EPR#%_ vtSVT #K$eތ,CRZj+WՓ&'N_t0J  XDP]]eH:W׳狻Ǖg^+p '@tws"u1V2J{sZʳ$ @`oYOin8jՆ9&fwi"D 2h|.LGQ+䚚oFKY {"1do&i5MBJH# iDi#_[@)v$“أLV@V\bH{|RdxHڛ`u5Nո[Y(eW/)I1CF$PF*-l;Nvκ*藥.E~*xg6U" ! v3}MZ$-WbvAs.ܢޥ2 (!hĴJ$;@%U%(Q_MCˢ?@(.:d0!XWs3>_:i%7Mn&[. 9_V$-7r۷)>~6)/ii:}d+JA&]L[&@5=dqoa)  J\ש؛^N%'cO*oA!AXp&|_kӰLbȎHܰbw"%GB (rgpL(c捨_'(Z ʞN9~_L(a=arΗocBdgLaͳqc[Tʪr񬾿VCK )ziP(k89ߎu"[ŪCda#BQw$ 2Br_xfuP^fb"SB8 rѝ[[C4qoúZ= rrhBY٬YNѡbkz ܘ4}]K3!0&J1-4MxY; 3ĵH+*< s/&85MjDP Ii ay>M~B! &6vUɱg!ڻ?- AʅAK~B%5̱i!2PA.f&qA @?A:w sc"_1EXKr9vȪ;Tͣ˅OB#G\]к4XmLLsѩpTLxr3^oKߊP9N;&ՠԢJ(SQE0 /oO:5Sŏ %*Bt{'{D"Ft5'e1⛽1=M_IЉ 4\o!@qHatj*ѡP}󛪍3eV+a4OTl. r3eqhsID7yS\-EV/qo5?%+{̮o3|Ey69ѫ; @2S 4ɛhuBijZS7q$0Nk2TZmضN&7ոґ yIɂ`vhNoVUFLu 3^-M7HI& W'տRnv))E|b%Sڊ|컵sRDyң=r8D55鎎DʲVRXDʠPz4+{Ed"/+̵Exnp㽍 WKh( /dd~q 717,Nx̆=SWSCa1طSxN^Vnix(M@|-kYoУ R>a9tǣPIKJHQj!ss-Td|srBB Z#{gtc]״2F0"a4qfVEY {&46QuJz'1@gJlj'=!oG8N8 3,LG;Yҳ*|J>-<ˆ3ɍ@ A63O ,zfC`XѮPF/CaeWi%E@gűߐH#^CKl9͊k1T⼜IEP4(&XU6`<4\"5 ^mj8baP0$k?fQ9OA)*9w)WhQQM)m g@!Mb3R1qף'ZCX!h4 wD2v6o5id_WᎵ8r5{r+DRtiad=> 8EHf+J~\0B(sėm / .(`L>g*qq]E;ᨁeVp(c٭͂aCH3ILHrw'5cmDp,{ A;z|"jgj!lR%U޵*ܮG8&rá+"p#„ rEwhna:&l̷hk$a/͘0.nwٛvHpdņ'.*!eF {8űݹg-/Ne5vFs+YZ% 4a^ vϩtM PwN QBݽmͰ֋ezk;-Mk,Kq?v^_e^K‰n˖-4VmY3 H AJjka!J'2Q$j7K&H$& B {ݎn|x]K,2ls]c1GJ,C%( e پ]֒T; ?|[Krh] nX*;_; |8Iu'>O[֎Nְ;-dPL\tebJO& oi:iKto^V}lXuaDO-ۀR1w` ;}IU -WR8h*X0dImoݓ.9^Bqc*Oz-ˇoud͘n)v;ݗQ21{iLg۟.+&]$l#g/51<1 at~>tVVFqA`6Q S8zʼn)ez$a=X[|Du=nPZ<<$Jc?˨ڇb#}b䂖h]-XbRPzu$-152F's>?-%qALh}Mk-dQJN]SdCq(X9?W/r&] 09f^jڒ.EHjsK!}NxGq0J+',3G#!-9NKB(כ(% _[a56#?>;lK}vvlCDkyaPlG9 >!/86v ~ܰl݊X-i x< YP"n6_YzG[]b7t/ЇI&W^NA|Gb8 vPllؒ[ABť M"YxӒKT1J.$:rOʦKzҥqjJW V:}ۿ^*ٝ}EOv֛[%!AfTWJ{'0+mMm"H!%% &ކ}1V4МZYOTXChĂ@8 *?}Zd\ygE3'zᆑ_0W.}Տ*BMTb qyWTlXu{*0-jIW\>Sb:!%"Pla< d۱.(ŒEoQŘk8JOF)Q(#{u0JRJK9tN[.dז9Tvx5[+`ajK#` ;}9}*ٍCi,2Z#%)V i\DGj=*\2FIT ۜ\v[-PNo?b*U_bԓY^q4?)ۿQ~ggRI(Bx6w$Y9DsGjp}ECQR`7aE:R(Tb NHQq3ִۣtfĆ8cϑyQ\SϙJ؏a!t%P'_&M 2\0q]~E'*;YÉHrf2~j]@Ox+ŮV* Jz c8eF1ҥጽިA'.V2 ΖZtأ5HpUȩ>lO0!}{gsZT H~g$@ +4ᾂ&V:wpC$Bl%Iv29@ZM@JqJ $yTRX1Y$TYH3YiؚUus .$7{SP0y0\mpx\Ù= S7衺/)t64i +/Ϻs{ɢ15Ѹ#%̀;Xl5BsJBlih)-,(搤ڨ9Wϵ.A:$c@ gw臛Ks5zQjLXGu- .ԓĻzv=2cB J b)ͯ_'IiQ硜qyz:z6g֒rУա.Rĭdje:)3"%0Cϒa2ѮE\:ZbTxxmEdzV\3h0U8/`R"J t1Hj؋pQU;qPU'# )V?W\7^Ob]R| (Y66@0""mY.>C4Nׯe^{ qy?~+^\VGު1K@*I (E*4bg,83l lsPdn1OjPѢ$[ɔdocg8p'mDaSz<`J7K6ƨBfzAo~5W{%!-].pMPIuetn00U5*cP]it,;2ZDCVy1Deo]cؓq#F r5\RI$H-ZZ#^ە=qa,PLLht\i'dHF5^j; J >2!8DwVڥX2)܉C1_վ>62n> n~ZGpXn yOϥS6Ghk xTtb:matE#ASP8񋧄VFiH47' ./AXHk-HfONʣ$;eLJ^S-t*ۓ.60p%2|n/ ޝP0l*SL0f݇ (Qq5?KGT<8ø Sw 6nZ1ID*"tYx=.0S$.]g1N5Ɯ+5\.V+'Ppu&*P;`;+ApD&WENL9z%#cAqeUXwhEIR&r+1Ջ\71;sL=wmm$ M}X$^ /'@v`GċLҢ Ԑ5حPr[5]]ׯ:_KpI<\tkJz & VW).*\ %[H߾R* t(P8̒/V 2t2ao'v:L E?=r?~og?+e̽"%ȇq:3}2V&~W헑4_U\a8 ^m* ~Nb8_G_i.o^ y<ǣz׊ KPttcC^x]n~ma$3+ m]/WFrS\J!w{ϫNcZeNZ+VjDۭg~z~ₚi*x]7&BG+¢/Zbm%'%y!.AFgn4&59hȎ"薛[C4Jbv? A<&\,쫞."4_ T)6E0+xAu(,6m:_ވ捊u {1 1mގ^ oM,BDc%.[vk39PbA0)8h4r"}{ iu<kVj횀 8,Hb9&wixiuu=&tא [%;4L"z*$ Ê 5аq#e5cu'i:h˩z RlΠO'wNӡ6\ u*CqpoP^("if(U`\ FJr !P.a)~1<3Ԋ kc?;o+xD?*١~~ #3\p#K'Epzn}8ªXZǽ:6ܴ(jS!0τG&s{xTa>W[{w0~>t=Cq(q:{/7e*)cv?uQ_(W ňE୺r?8pw\*x5QI5x\*&OPRˮq ^Zbd]ѹeqt3DQ"TYkSvyZQ]3GgAkL%\yB9W1 O(R8 ^)u_8l7-GWuI)5nW>< KrkåHۆ߾ V8i2}:if 1#ss /EOyPCYRV| % %>/1R ǑV#o0J#(G>s)p޿` p7pά+ۇq87UBJnFkݕ FBssڮPU1Y 5E/S/5[R|SZ OǤ󯾰jןU>Ynӑ_cnqrd?&ROmS<`nӧ1ɖ?%qrlK楋 Od')_Dg5ںMF \901CgYOF|v.Vy}X;\uU쬓ߥE(QO{֯o.rkU36Nݾ>m,BP!ރ5-1q笂GO4c Lrde0<dXvYu##n*p-X\Kxޥƞΐ5MQd1Nl< 5zUa&CKa;r>[v6bwXlZ/҆{{v5^eZT8.V =d"K=ȷg:]jʉAgȸҲfYC_v .ȭ, B7]6 pIh *~I/oAqY{Ow=ʍөH͆1U^X7io `H{Ep7S'C<\;\Z'm]|D4:௪'xcNYEPŎc9&xu?n^T >~m'L}0G?)t|tzOUr8 LZ&1%ZnA豱!mJl:]Ymx&>7r0IbƂex3]Sw5"!gOPj%;n*k +%*9a~~}7@ q"Qs7oDD}HTLgYcRK8 8pj3:X@l;W(ˮ]:xm-ٰ%9T3 -K&LP!.3AUIm>>Bu..7kmQLuUdGȎTkTh z$y$ӎ i^Jfȧ;kd){TVe6'pṓZR>ܔnd[n4{ Pd'ya߆ӰaE^a7__톝[!;CZR ȑv>~&I[) d Mq.-{E v,~~ʍL:*xDC^XBT ap]sd(2ΔS):@.Հ .C7#JB aYZ5+0\Slk7 +͎tcGnu1F۴VO,>>gmɦn̐xV{d2QwKpR%W(f_i^6CFW(_Or]& gy y2 "+5 P>JJbuA>P r`%9"E=w{M=E Pl*xQVt&{ɀK=E5]D,tah}rPD" JW8T9SJv~M4"ԋSr&t67L IaYn~KA Y-bؐL(lP>#x~smf*A$.|> >EkøA]hVKL CeA(v6g*0 U8dHt٢@x 6E >0!6P)OylUa4ˈ?]>5*YדnKZC^Pe4)ݙucZݢ]Al丵mI"zn R؇;QP ˶ |>%bMׅ yzR+;VJVq[x4>hȾ><>ݠ Y5d H4KÚ"]@oUUWg^ /Pbj~^'&Z,ez{̼4A#q}q n8F$H'h`f_y% B!|&\9WQސ[~CFa •C;pάU'TK܁jpT\[^"A3U. !+gyד 5jHܪh9-B46D' \ 4ޱF~?M70_A갂iGۊ >1KРSH lMgmo=їGׂı nȅ{ +x2!0'.YD61(^Q ЮU7>9 98G&k'z&}cS3Y4!eEBtqMkRO2۹py;%P [_g&m2~Vlӳ(#G% f*Ӓ?)A2<ή?FUױYDb`ʟ/{28 q0oZv;`%LC  ,zLJ9K`ئ%Km$ {*bA782?<Owl?K`PpÄb*&+ vhD3/JFLzORs1ڋx w#f?O>OxT:td1^V ݋m_w!xOd7姛qϓO5Y}LB2H-N*8d'+OGA|Etrf0/7.B 4Ȭ"d:o>ѼG'r qSh2UHE>Kֆq##\*ڰlb<0L;Kk*N)W ~e,CI "k ? 9ȹC ;: 4|/_& D꺀_) u.'}{ϩ.uf6d_$m-c>Y 6hbKx2`4"㆚KAS>T^[A[U-\!_A/m a&(#M?ͦQ7`Y,"E5U l*[XH SnԚGhD9fEa.8N ud!7]1 ÅøE{Rׂa-.6d!\Qƪi\PX6i3 Ηl:vSFm2ϺMٻ޺q$W+*`=yX,޶c3_t.)Kr||hqYbWŪ%+ZNũSN avB;M {{rsoE'8*q 󑥀7UKptHJsQsk}g`%\~shzUxzl\2?n&]Xʃ8&w M߀bj6hv3hboA Lgo> z%S< f645^'pӹggWS?Ύ\j#*][AMη_}4u@۠D)2']Z AcbǨ  c~e[7:7qPsʞ{~B_X3ߎY|=2viK7?]|}f%zddc^Hm)a4Fzຜ`s]N u`sK#$"n~{.d ^_q}gj/]V];廃9#n.*"OUd2 ~g7m]efͳj\1!نo}9q0/5I(S#cHzS||XՏN+qxR#~_ڴ[F })ꌿՆ/.y|fhrr4'r8cn.&W›*%SSrnSrʃUН|e{oC cuy#r!|bIV&j[ v^{+yvXI>l` Hsj8k]N)MO?|L|C()gAbbq6 j++d9uStNVCbkAdsv`H &ZH”" *|[SjQUP--7:>r=>U1r̭(/}T#H zD+XUa 9]ȸX݊~n4L]\ty2^¸yo5Jk;(kzp0M!5̬p4 V9 ;y*JVV4`⨹:?C"8)ɘgfMy{d ДHy!ǂ`D(rjɁqn=v+ݜ !abo]C{W?82F#Hrѩ)i#c'---s\W5Z)|-L8iUm+l} (9lU,a}CLׯsw1#,>xn:Uo3F 4͍1RY۝#/gU@|bĮ"աb% ~uP(<\""{8L9h ;Rt{Z?2/qC|6VpЙJg2kpGe+Xjڕzx/AaH£C<*ung#C03-z&CdۅINZg8dqdgdxF4żL9a γ/,sSf=' J`wMn&&9둱52CT4q|J:RU#c{j}H 5{tm4Jʆ!RJ}hʹۆQKOOTneM Q j3w[ͧ!JQTW sBGR{Szd :S+OR\gV o=FUn*О|箢+j35R4W[K sVDV3qzظXKvQF(kظM)EȢ3 bQn6Yh QRdQ,67C.3Ws /ޗ'Eomrlyj$ 9+2A̖DIg7_Ѡ:Uqq=\/Z}ͻ"\u{ q%ڡJg7/r  δ 5݂WѬ%i_7m;a-R!P:b*SBvڿr\-9%i?_5׬>QcwHde墷YQy DHyh{?ӿռ'p&_뫚RXZ!&FIZĂ$Z:B$1b#t]Foޒa[d-l();n_G- q2 j⨽VY(լU>P])2аN3בbB`l*F\mypFRpey [: 6Y+@QJto@jX+v€I$ېD!dh4vPvB;hBZMF8nPhhW]3 hZQ|mvfN9!kg:P]v JnۡQuN[b"9Pv=b4@WXtvfN13fJj +n}C@-vv`#;4hd@MBPvڵ%% [6swE)6@ +n}SZuu/_?ۻ YߖBsu腧][2*|?m{ .pV{UmxeL!+/:ŗA@achv$] lKT4Ldhe] Dr֮淅vދa\җ w-{czZSkB}OW /qg tJ0$@R0'iQ)>'LB});siZuX=3 gټ9%PO%ܡ`r}q ڽ&h7.Z]vF&hoج) {%Ў\iZ1swŀn)H),Ю@ M+fiѻHn4ǹmHRctGSYMpELi%JTJ"Tdՠ`Aeх /(z- MX[g\66Dw(DA;oEQZ cRs75gXo֦JK#eIkb䢲§B˕&t,Ru^՚7fSQ~SWn/b"j/b"^fڍt[:~yH?3WAHNԴq5防,*H5 GNs*S):䭈ސ2&UrvX k8g,@BwwuxV1h)tK@OkAe z 'q)kk6 YJ&8b#x3(K$h@!bi^ VwE+ހ#dߐ5]zt=Ajlٳv&AWl wo(5r$hz0$5vq=?rX>#Q?1U 7z&R$HsU`}KSnTJԀIM@h)SF%S~XA c42BvpIavl2Dzܿk>73I8Cۻg/jPn{`c5ܠl6Jr Y*q#i ׂ\Fʜ+Y(͜Sٖ?@]77ՖdȎ@++w Mm&m+(d;0ङ-gy(:x({ă#uVB1΁G1hk ]Ȼ"2k%r_h)4-%G훔B9o(Zpj.]#,Ĕ`Kjḹ/;Qo]au~R1$% xa@;jkw) 0b`jIJP€>3NϮ\ 5~5Gl{-> IcU[ndMAj`jQTI|@E1Y>y$j#RNʭ\+hTXwdԤNS[IUz:}تJ{R*\-YWןEYIJhRX$3h *|4 ^ڷ\JԇgC#qK֓ay U5:%(hGf߽MS %%>Ozl~2i:B^B hr~4|ܜ 'zsiwe\@RtbP/ ~ug"Ă}Jlx Zyha> Xe[,;$ q>[Q[s+rd21j)m\\F(g[f.SI2j+Ӗb:x$>'A9 jqN&Z?A#0)q"޴(HAGvtd¸SxPo-1bS]4N[Z 4&NqdxiQ=ӢzeZ̴l&GדWSy=IlǷws 9@&uz|Q$PIFOŠ6;t98ޮQ4Q(Q<T_|hKn)QUHJ)/dMnIYMq k%ڂ%z#xld|IaA7neY&":0ړ a6_Z~.K P'W 29 O@e>EL66q:A!nsı/6\P% W~sTc4Pbv@sb@8P.%E(⟣BrNyS$2W-c:y+ްQxi$,9uٙtI{`j ,1f&~C x5mqQJ_ (n1(tuқ\SYz6;tZ u%+}>(B %5J.7gMqVV)\n}fw* ُ7ZS:n+{fR;Y%АwX5kMqbq~J YI ˆqc4Ľj%uGW(BzzLE <;v΅:7z aDɞr6tHΎfMx0C@Yl|P܊H,8v=>.)8[/S:1Ռle}Yq5SjYZLȬۼ#}ao,[o"_r^ $N:ry|u'R(^}B֬ք 3`S>O-[*Va*!%[*ڎ=({9fTI*+0<*u&f5]l7?iP\큛 ܱoii{̏nj:v:t:`hܚϮЉ<!N@̆>zΪ nޝq~1V'l5yc elAyG|hCbB6"-y:CBr\p6/4cŷh =YS;*/nޝ\ϟ/&`u Q/Q'VѯRF"Ɍ:~k-(^ lk_vN-K-@O\Oo7ge!q% _2hD;WJ77~x)1XXaefLo75]D­E^v@C#c=_z\Y#+>hq4PMW;'2a i~a~W]`lb6qv;P}3{嗿Ϗ{l,K˦eկO(jӷsq* Oq\涓ǩJl~7t|KMng]n۰eO!f=hCq;./~z{7S%γTٚ9G'V_`0KEkyҡEޟDQ0e>0bBWhn} q`yX6m\r{ܻD91iv׹gS~q9c0Hb>*,Ag*ebڒAJQtt tܦ9jF+&S%=bL0dLp>k2bbNP81^5K{(4Ș,,ͯ{=AuNrǹɻYyR8ŶMtO/V˿}'3.Mztz3 LN<aqQ>dfyCvk*0Ȅ Ս]o·ű'`0/*v3X!Tb@_3 L)HCpXN&\#hӒMJ!afB%c5 [B;_c0ٔX2%KRd}Z7މ:ϐ%.EZzY(Fe8 -GĊ؄rTWRM}*ɘB!'<86ЯjM1 " q{sM/>pa#Wkrbë~$(K4T#'X܊z(C_GBc`ṿt|b !3*tZ| rmv{vNQ}wNQ}sFЁ{0vШ|Ej-n5ب{J`/7g ɿqVJH=++\!Qت˦uEEg6u0֓ z%>;琚/%YTc"L(=\ 9mdkfs!PƮ5 Y_օ sJDH* ,:|sQ)פ ,P\Z|ٖ,ZBAt(h{-K6L 7 y,5^Џ SF?O>~Zbt,$@Ԑ,\\1Cv:BpܺK+;o\Fn|Kjix6h@hs9e \#q Ul[M]Ōba١!h糛4_}A[_(DH(xI;XhKR5NUMeVlØi$EОY5ym_DNXI3rb!_$R@AA6`v6f1֕~N$1qw+uv~ӅƂD!"kSI0\U1Fd$j% h1pes|UAql~߾ehcPP#}MP{4~]16fB7>;$K},=̤d g˒zb)hV8 rdEA9KbF# J6-"P:4s󤅞j*U dzCJRbӸG8Mo<%{ Y~P͒f£tƞUMP%$6J7{,[ǐ>`8z}H 8pE5.{}fæJL FlU-/e1=\وaGvX $lޮ%eN9JB\k8/Цx7( w20uT4g<6SBqPj8ΦP3?=Otu ?mK_!iFD&yteb%]/eɄ,RtTu+9(RUb8Swv; 0"?]|c,DEǟ7 [ԼDvӍ#Z80(D> s5֞mUcE|դ=eYT2Ey>ҰK5l1bbU8] LK*@[j]Gg$VGVYse/=J&b  ƎجjL#Jۡ&[dBlQ*GӵH瘹$Yn4v^emKGe/Yc.䈶8/OsFjT\3# 0~ٶ;\߫_^nf>4dZ޼l(ta&Hk&A{Y{K{9jh)l"F8i%K]@.cFh<XOX}M} m^\ FW u-$2p!VӺћj#c Y˾L$/5Nѥ!wծ9Τ—gR{%(Q+O]ǑS Cb.@fƝ(a|s&r:S>sŅL<.ZIHkPr$)t|F%VSf( 9@3񊒰d/ţlEn̄yo/^3a ɛK b@S]bOxuF&qM`c MreT-s%ఏ۰8 Rxw;#oߺyx'rRB]n, cMzZ^b8aػE\{_;[ ~2yOS #Gw;xcPaKlj<4.̴=-ƦQ؟SZڈ}۱ a4ܳL9ӴVr~d_JHG!uӕS1<0F ΐN|N}P_TB^<9κŠGZ#ЊxMs{N1#~o/~pNެ218a)wN6h_o>K0v7?l&?iV˨)!.@MMZͩ7@vSNdeE5--{VWhj)vtO3ѫs;b# Nxf slʄ}isO"^;i$u5D{Uע^SNPԜD(."!5’I e6QMٳ^ s vD2" ܚcYWgg}yȭF1nitAϓ>7\gAoly|PDU_6C5eRu)8BR<nwг9הSP7Jd!gH;6;EcFkl όRϒ&)%M&zTIzYy/xʲ7΁8e?n½ׅXtQ3MTvOԇ"%EBزϵVbdz#p)&w`lzʤcSoKQϐh*Seg DPfMxR*y9¦.@g(O NZk6zސք/YH9قrD̼+<1i'MRpePhۥF5%YIw[ FFcV+\RR-\*, gG9t'Ո>qj_[퉷FdaaBs*F1*B'OUG?O2dOsa;:ov!~&箍ח/]kcR[oG}{ Oj|kG6>ϫXp r&%qwq\qvxn|{67v_ CiEUq#DsļD*l˲qvm$w}Ɍimr!L}:[2;k7r">}|)E7^<M;'6bvC@ü{< 2>tГa ȔSXNm 0H.J-Bid[t]/P6|ao@|ȱzx?X޽=jJk o7r_?5{'xeU#]\޲Q~1(eO>)"w~q7۱>xѱ'nBF׆'5&^ '>pdO'sbTB KvKo#>Syc^ڂOs\GX|0ŭ昹>hgmIm (-TG;mѶPfK Wt87`N*vl* Pfjƫvc ڽHjE(֯F7E=!ϛgFdL /|-g| V%A&aVycv<<1qcצ^4"&Dhw+{*gGu(7r}x2HOWQD3_DaX:('jK@BQHm+8 XZP0{h Rix%yw;}uO6LzJAV:S3U` TgfhJ=_s!.̒bًӓ/Y2ֹi .[و) !ν3[=VND1_C0L˛wuqN?.1\%zc!"OZu:,h9w|ZXX{Y{&Tiy<(Y7HM{1&7dsAq;HFZ8pS7c}J + >c)O.tc:hqrma Ļӟo|1Ƀ@1p>1I^Aor`~p$.&?Su/uO4yݵUԬ6ye.Pͤv I>OjGfSUFoGY }"ku'{j'{C87FnTI dGN{Cv}n_|~6zRD+m#IbQr^q5ty4uز(Y#nw(*U|.VZķp7fFR5Ԃ*fy' CjB^Vu"-\*BUۖ˜ H#Pْ2:Z2JM, XYJ{ mX9`j1_1/nfͼ"ֵ(l8ƢmjUDĶڵXFkCc+ThUqu]:H84Uٶ}}QK[L!(Gc&vIDdqx7a/Wyf[橜e[s@nukz==pF`0;#sTx:=;9ߘl;#O=tOb¤̾-S5}~Lr]mY]>ةqa /Gz`Fq3+'n~1K| Fқ[d~a1F)FٚP.q%K*痏*7g-_w;r~ˊfcaIx} ~ԌQGn\W$9)90'uL#9Lq<#I;<^Pzs`ܦt{xfm0&K< 96ScZj);JW$>,tb6-o09:2&cgx36uÀA?1=i#uٟ(ULjw0ŨہC_D*RŘT@0g5eqH0u~/{sJX 1J#n JTATϥIJZҔ VB֠>, 76P{1HhhBﺭb[W(i*APۛ0E=+[Uuv'"%{sZKhUI`g0D8̹qj O".G;F|.l| *T ۄ$U9B orGCR+5k#7Cȼ*j;v.Y!P-Uw44sEΆ4 f6k>Rfcv8iǖ}ډE`xݲRҐ0K>[c^H('cPӜ]J f!V6rߝ[$iVfq3Ozorȋ($lap=8dQt Z;CG  d&yA5Vb,nf7;eһ)88eЫ@rZypdXzMv<XNGZ:X% Ӧf+sB^;Rs[(ÇHL'ꊽRй.ק\]xa_4x;_&BoI0`ɠ7.l9fR\D`qpΗUM$Ł]Hg)ZbC^)t[ұ7RhwKs`1x[t7\ x".QEL/ m,hdSnI氀]&Q尀#L Sk  fss4gFPŹ;2LTOg"܆c;{i*1;2Z%@ Z'LWM{ȵ#fK7~mhs 9 ;A&r8"˧s΃#g`]; ?7^!|~3ҹeTA[Ol L1q̋  @QLYQr`N\\1vù7?Cj,t漽xA.ȈmQkQ|@e`#+:`W.hڟx}nZf2"}eʦlՍ/5]rh)ziE;LX aP?Otm,9BP>LbBQr_T #'e?sz^_,V\3umS_%-*(mUvg` mWհ64*ԭW6[Qhb0 VRڈU,l]M-7fGsރw,OMiPG#Zzf&7\ y&E22q'>ПgI]*tP90>@h3b~P~L5m.6`RiR[iK#4QS9Ӗ6K;{z2 KSyiB\ňQDe躭""VM,[+#lغZX6?RBӰK *QTV,[ޏINIZ5;T@AWE+.[&EU6d[WPTUTD֕QMIP87щlxOQcت,IQGCtybe ŋ`k_BVJSqlU nKB𮬔kCzt>%*h#r{8PiA2R"_M`MV59 $&ERǪd0Z#.RD5MPF-Xq%cI,*}IS: MS_ [=-瞷nCgGE.-Iz81Vl>Nه`ȹ aYJjp?+{z{ ;Z$AF5w[>_{uCűx܉ 1"3D~s* AXסdΔ-=NP$)-bYӽf-"\OM?AqLAZ6/6X\1_V"^'ٗ7T/) Efm"*B[6)q]S*5{nSsbo~ 12ã ;$=[^w3<7׈pà%۞7Y?! wWbTyɿT>δ|+:`Qi,SiR!`C0ߘ*˿h9F;C4x Ud|"5=4 586}fQEb X72(ٟS${bnYrݻCYWk*~I"I? s4ŻfVN?!{MT {.xrחF]S].f>v={NϨe(ːt MwH_7qLF'pq̋1p#lac8Ij(ٟ3%yvcN|P>GpYW< *0yH9?o|R?LYt;s4˚h1x,9eP`e[vB'e1iHL P̦՗ u5imM;@bM;?7JԢ zVg&wn+iq=)Wbv#o)f8t"J+&`pC=SvxEް#=n !/=@TNNs+F<'1gp"É-/tuf8mk =ޝ'0p!F,`Od<1ߘ p& Fpmu|b"KX혤zB`%w蔚l0p/0<{ȎO 0x]uD!0 `Uxk;9TOEpU ٫b!7ZwU1Ho).Or_VD"u]Tm~}íZz 늫USSS'm&Wo_;@ā0)0aa?>_g|qt 54Dl!ٶ-uu,I1e!Ӵx*ohL F'>j=Sjۺ/X:AcboKqC*XI9MK1oA^?# B%PQqy^0XۊshtV 4umj:hQu谆*|pol#e-ك7I3}w!b sQWl#Y΍)bYr >vM;(&v35ؖX@@(j2BsKdJ6 JGQ** #FK m%u NBhƫvǵ*C(ki#GE1E Kwև8BQxie=F+R:ߝD]H(4!Hd~_"ID_|uɾS?F%cFuE#ª(qs'9*t7hV(@գӲG4rSkH>s>Pz3a ܽы6KíIYwVݓ87+VFLj1=S[s1V5vApliӈ1`c\rPP/CCo‹:/G NyO MhaW zNhGNmq*&wf䜑LJfM>8PM?e,BFCoo|AukA#{ܪy2-K'r/g|:>9}o)}l0~1ٯզ6v:2̯ȿsU/ 緿O}re992&'PݨMS&.!+_hżrBxޞ[/>F_Dt%Nf;56`ik GԎP͙TZ%%&CiyMͿ\j",4K)7hW;>hv}AC r=wF)ax{Ǵ݉3 48 Ȉamci]\y??~ qޫn84ıPjF;$&co98oR6iV]TjrG9;]algE΢Ugq(|Ӧdq@=I zy@oAoɨ//sb6ĴU+%: i"N32)xV Zh_S gCs&O ̻4".[>dWx6oM}#lGkV1x}oΏ$DmotGAI'M'G"xvt(sFͭ8al~&jH!FLw3նJh9e8ٞ io۔#FDz0mT\OgB FjDx7:T=vNjӨ><'N*(’3tFIo2} I>?$0;hą2Ԙ|Hj`~ BrU4qO>KUSgJ`ABm!cq("On2lKeIeՐj`Tr߂^m \Dj2{'֢Cfq'85٣<ج[!^…ۓ)-=q SOhzB㴙и֦[xm18xU &nW=Fimپ.-մ̶42۴fBt29>$]b ŷn!0 Ş /|BZ9$fw rI0o __Xs'ϕS 3"+ 7$pOu^ BD@׷~&hϴ̪Q™B~ոG 'g9f3ȗ wLف͠ۊ!{淺U#FIuwK,5'e ^MREW9ʞ*9`0rHsoƪ\M]U|g-*x~ T/)뾘0Q){=7֑CK+`&yu?LΗ-Qݢa6mrߺYfRٷ萩9DHF[#"z|Y?o&Mmgb Ga 'mʤ"yܿ$5ONp J7Ω9sr 19Q]L]l,=/ m/]qͩVgW/_zwbPĝJs7[o/\ۗӞ}5}yX ,E.^o;|? Y1x@؛qĘ|2sˤl*&>iI%{ H1zb/3\Q<#FKmZVG4s>r[`6@5O{s[M_8@Iޓ#+ 7nq[b# }<)SJl?eG|kkTEgFM9UjIhꌥְ9lU+)x/yKǭ2D&ٞc^'-MMjň[6K-=ٹAC# IdqvŢuɗ5Ό띆o!s]} akL {SSO}?xu!Ʌ{8WXG4rw`\{y͓䩇=/l2 pNqstꧩ#R>?R>F@{ė< KA:u&HXoGp [^n^n,f][m ed?zu@:KS h݈MY [6!%ۖ3׹0DFA8+ LvTZt9R,}Z1YVXwHZ`ohzuIq,_y4z0=ٞ%sh=wm9uv`1/r32v. >Ae-I-3׽; BǩAe C;1svpwDA;ቓ: -=MXq@-3#Gc.@vo޼( d fѦQ0<'dZuqyF<v7X>B~TKvTuqU@  z3S t hw{#J( !M=rҪ}QGʜ,V:7TT'9DUhX9Oq%&C-f ]?#N+1$=E:drHn!]5yCr!ڼCw4;Of XUab_ K\13B@f2[s0;# 1 ebL?} c}}LgI}L kh<56KOKi$jR6v . nsm%hIBsͮ4O+>d3Lέ7[E5`]g7TT0 u9U"p8݆wo<A3j`;blz\$J%LLDAhpl@ZcawXZuK`+$[G\+juZj3 :TͪdPŰgq U(<++Ax7Wo ɯE$MЊ$(/@^qD^֦}z=ОO2>QN'aY)ԜLX `{R1лT!iܜ's.!"jXI݆VXC_ 2:wS}k\TEи[ A@ HrJÞ֬}(I h &`3ڻKYH1ޑh:'ܩ0)yn]!T-r$[55Tb[B8({Qrw%j6vЦ0ObssU/m/! ..6l|->F}.*){9^ " qiE<]|7o T >D )&"Os 2]I^pش9UtȆ`|BB¨bPvNsar2g1|QKX#Dq3*=RgM#N^F{.3+5Rn2}r;Iwp][@'IK'{\>4|p)ݣ0M]mcinqgZ2OwUE.Af8@nbH4_GғWtܶ,nȲDVUz*xrqnVR(&ءlPCΙ끒ˡ(zo,j:6oSP4.ZXPS\E^pabbzU*mKvjmbƌ> ݸ/UtOd^u537r7TBغPE!Ym/CBߪi,Q(jFnm4H RGZ?o*Fիh~+Yo[tԌ\>N-kc`9@ǎG:ssiE[usO34qkvOBJv=6&Wqv~|CFH}2?;~BϿ;> CUulQZe,+SJʦ@bj& A׍w64uD6l+u1kԱnh(@YּU.D y&y_:#t}jka(d3%[q*!M_e'Y lz!u>^Qc6j;tmP_[}-Qzi&hinmW͠# 5PBM ԌsUm ?jGul<ڽs2Yr 4#dgn%P5{tٗ:=c`{Թ6Kn`^vPŢhTWnly49fFr `rHl6k$Dcĭ{uYOGg6Vgkb6UuFX8!#6CM':v<5>s:}x\<7o:P=[vւP;ZAjD*,inڔ,I~J9)ffZr 60`epy~Y'+?tB=EcNbG2v?;"SrdOԞ{"xS#? (W͞Yvrl=sA@(Sle Kyv:Y iFm-ZpDYyДņiHG 47\Ы"|ayK ranz`ZJxDq =;UkuR f3t?P*LpP9K("<#EmX$[ ].zk![;b*;([ϥ(U tX N_0 Vh|2wd'h!!;3v‹)lXt<\av-}Af3dhΧXw>?'Pb:2<$ifdwhN:N=SY;eNlFv-}Q|Fv;:$#b)UئV b.X M YNB;)Ac),ex, \v>1ojۺ  6Zںd1+6eӨ^bTۺ ymGc :hS3~J6VT\4AH$~È{ .Nō6l2eƣ#rM&@Xcveo1D-`ܻc'"oP6*M]$QcI>q&XEa𸤥Y ч|wT^LĽ&fM^6w_ο#my\Q1>Fǫ <ۇU!kq4wlg?I}P;dF̅lPvaVmoS")\FV3| FAOtO]E}c{6b#ZT+4kc%@j:,cUS1q$sy!hlVadm+nkor#=RTa@꾜:-{JNr~}祫Wi{Uԓ˗yKɳ"Q[2Iʺj!+HP'Ϟ;5ɝײK6͞deL^b훺?==)eͳ1ndz^OhWmHәƓKεUA02II\r #nW]o4aVfc&e%"T5JzrKdCUA Ĕ:.> y#Gc%1%?O +y,RhN)rW\F"f\|_/9f ߜݭo<H6vrw&? ai{IW²q/ơ*G11ȷA*;L r~\'n'}c.wܘSj3zsmÕ9\!8{/ٝ)m&dzolvFe}Sg6Jb-VwuN2(SܪC^y7P vkP (xH0N YFA1p^p9~g1nC VqU>C+DǙ8co<'RE_Gr ldHXx#53JsJTltƖ8A*PF.}̢:3#Avʎ ͉9Ph LȬ2ЮC #9T؊ Ya(VHov+D9mgCu /#"|^6dllld{@Y±!ECTrlde ^蚭?m"~ Tz~33wOkFϋ\(z~򶘜뢝Pؠ'sORb-pcZ}"1>$ј>7ޟ /֧@ , ES8|]- dAjhZ=7sP,#XKopmM%UU˦J Œw<`)d!|t-[c]ǮNI(jxQTNXY kfs-œKYlT9CGv9gPdX7gUO^2¥D^t#=֧g_NFL!yؒ'v FWw`=MUa[G%(lB5D]1@ /[3,YyXR2'&nCzĽ -i+]qݙi^pX{ɮpH!r3N5ET"!U"N< `z8R)P02G{ c M™ Ep` lX[zgl@z<{H<\ i-aԒwv{^|&XYtĆŰOP%!̺YBy&]<$z!񥔉22e;'2Y͛ 3-9{bJ^c3d=1Hڥ!2!ڵ#1bpgNk4iCI l+UW !M-Ĉ9xzeVF<5-gdhO9k:^gi-QR41WQij%X 5]3Fn.OmnAlD4;fzBBڇ0s/ JY}^`wH/v4ޑ:xh@8k&__t6B"W͗Yqq{{vyt?>(2Bd0U,\KR GSW@5QJHӻxNf'T'ME6M\d Cr16UҶU|#ȁ`,GmbL掸J%A 9<[l(@N \ ؋XB.ka,Gh&:,} * Ħnkxji(r :$gjhN0E}Nd:p@rF`ъ(@݅G8lrԍ+Z$-KQ9"YP!}QG[vP:S7, iMd(B Įˆ%wf 4cIoyT.1xTx̼*l \?˜EE5iLnj%:ZnT2tu~Ju⷟mִSsp=V뷧{ٻ7n$wqr dφlCM6=<"i؁+vfZ[T$AV_z&pZ_WϾ7.zOL\5Ka| '@÷8+z]?{Ff2^b{[i0Uc EX: t gޓq{HNSL*Ox[S7+)Eݒ7zs'aO [=#Ŷ?YyT4'-Oh=Ai w&i_ et}0͐s+ze~yˋG3> :it^o2G0lRa', c)S˜"O׳= BslWKld*gP0cz/˔ p8!֫>FJQgTL%vH~l,ǠJ!ne %NSfc~/ })5ISyj0^yQL䞞GTDQJNK:m=7ecuşwe Lӌw{0 u]fDji@ᗒz )A^($*H LhʙWn='`6=EWBSv !w &XcЩ)R>šY2XmKD׫ˌKi=uՏo8QStE4OJjVO?UϚ[-ϥ:ϰjݸ,ibD<̥.pY$U(KZϱgR R [PY q0J+ Z+MZ܌ۺQFJz(s R'5ݙ> 냆kI{}>uʟ_JBx6r5`Q ~!;1ȎA@v Ȏv<ɸU eQ96Ͻs $ Es-9b-M0w @_/,T(/*!G/*wC,$@H](f_;Kˠt_͝ CQm:.TT1,o U>yftBfl?_L3ViȚL̮b(Cb(ڵ.6MŦ߾شOŦWbwU˵[Y8f&z K3. 4K4JP _"<B9i /lΕyNX8r/M{G|7|v~9%[xLbmZ um+Twjou񕹍#³ĒwGE֌"w0ɥִ YUuGú Q0G.fB+a)]sn⿖~rNc9*I6eUOξ?ƶLQ8T>)i8|`O[Nk1|\Qm+Bi }->˛ȯ$JM"(Dg aFW;xث*Ӳ+nШ4f8@UW7 ІǴf>+}Pv-䭔a ASyસzno~~WL{f2~*4(ߌ?7B9[>P~aP:I  ZL $VLqN3nz mɝ߄߳9Z(mOy4%e yhThb%h zb<;<[K8IsU0!9V3d`I$j BҽQ@Y#+ϵ 2y%8{&'G^({&`q*cR0e18j4| Lt! FOɠ-pQaS4G?Q Zg pn \p\(&XyTm8Cb{ b o}Q$+6Z Wo:Ɓh m֗_dGPA=,38 (LU>N2aLk;5;!ъØHBk´$Q&7"Ix8~Oɗb%.]#\VY3/c™ڿ蹡xMNX&w夐2;JMu~Ȱ"<z+˧ntMaAoݻgܓgf>Άɏ<.wIX>? ?-Pld](zmJ@;Z6-z,?1E=uI(@>43;1|9cB+_c$qI4>:Ѽ=͛蕎( X IVnoSPo3&SG˽nI̥T>,?$VvN+`[,lǻX߶+hxCYՒw}hpoSR`Osx~*܀2gv<~ 22Q!Ґ&wbuv,2IUa r"/pc VtaRVJc zP^y =ؕLM=̢ m)4>{ϧemsdBG8c%D3rbL`>v(S_HrBjFs W1J!|w #r"ݦƠBB}k1@ ?*/(-*"Frq+GsØy +,5zFrnUбi"\\yމ^Q V\Gkbd6I:X~w|7yv[OQaru(6Եk\KXëyuOYM~toelx2UӛhxEYCi32M&*y4?DV]fLcpާk|{J.?5iy[`=\e ,4:䥖HaVXu[eF~lXV G%[R?fx#D-˥ uRHq,P w8~35OԞgu\ sb,XuZ_( i2?}E Z#NP*rWyxid4~;L>*~A(A(4˲&HX<+ oM.[r@ZwD˰KͻEx|S&u\^J|nc`V@2-ٛ^袡K~֢BHez(P 4MfS% m}NM!?[P.A;%Kϱ4\ErꋒD9y Cm4 Jqt§nC;|xR3n3̝wzApMw,c㹑Xe.O rP*#JXAkb b* S_S_R(~~L $)r_'p9Fq KqaqrLDY)oGF.6Ð(miL&uK{C Ҝx{Xޜ舡b.K"cf), mF $N0D 0Sgǘ?{W6{Ϟ&|1m`ۢŶ(JulvڴEoc'e'NEdY"3ꀅqL.3@ގ5К+m6h[Xtgh%$88`LgQ<˨in3aڈܰ1dѬ0YYJa]O<C~İSImz8W)6x}t섢Ե"S,UN@2YN-0E<o \+μNj,\34a-M˕ R?=rebBbd%ۄKz\BUR@Q%*`' k6{ե[IP\ukw|[!P)l,MmbU=KjY*z@%^`-\g{+#W$B}1]'=-xz{SV)}ӝ4{/ejUuB+.ZdMB46! ѵhH%=Pyɷvc]>+W/?{og+O*&}^>W3V N}^یnV$b0JԳ 2`rjN{뀦wh@Jnh@SV]@#8RS3nn1tnC71TI)4c(; V(t#fZ;LvKgU+ҕJnИT535DTYm։XMYh(df Iͨ!>7_֕wV`YEb1aBິc4nnIԛ4HI*vѪˬGn|̭3$1V;+n4xfgt J˳tMioq xVf ֥26rP*[!7!C;p95Z95՛5%X 6? M J|V%zoa#(uS9:gNWfxǘV|ZNiz)mr)}>V؊$=})3{a?>/5(^!B5nE`Bs6(A94sf*htۓf"^?k-L;Tww\DJr:^eSoڦfjde[Kvfl7?!OXMnjK`CD7oӰd -Gj喫f_ d&D1ěLm2T"r*U1T.1 QEnKӼK4{7Vj"''p!uIܭO(h,nh15 Bۊ~ڰ9Ҋpvۇ5 x!Fyr4_֢haNU궀ۮԝ%A$Xu()5KUnj *f .NJ8+Joys#ݍC̕rg \1{OzSrABգy8_L>ϣ j?{0nѨs;\XCïe0W`! )fv*]Xj =ܤZ׸N#|?v)i>Z6#L\\|0_c6a?t<ɳ$[+ 0"%4YˈJ`o*ا} +Gp%uҩ(Y2r>P Cͫf& NEьLw𲽁Vw;sh Ѝ"\RL޸ <160fR}z5E9FdLxb"g1f7Y rgV2re7rm{[DY*yXucܫfpdh('/v-n;O *'0e&ݞH%~kIXpD{,ktIHޭ54,t|~W~vW>,, 2܍^Lž4&R(Oi#im0^nh{ЉՄMESkE~z'ÁVk=i==b#Ǯ?Z;|^mmѸ팊ɅuIp97<6M⭎qԽɧIw_|lGDׂ{DP Tx?zg N3.Ue^\{1CȦq ۛhD㨓q'#8|:<BY&}ȓQkq~r .:ÁF_v;w/G`F?y||qq81?# U(ĺ'^dũtg߾&o Ga~xɕMZ0u{<;;W ox}}{S?q ʰpqR샠U`aJ) _Pu`),(}>V3J`- "$ԧWIiYo'oDC:biМ}[Cy&4CTb8 \id2I%4>2_aj(4d\ҜLQeD#I+~T~{r$"_ *%^ӆ7\~ۻQ.p!7 '?ZCqN0 q 2:0\dC&a1 _Ny\2E_6RRWIX5+ lXd"Mv"YaE@Nׅw)`eBɴ6rTPcޑYhN) AYjFcAOTRf_"//e"ٰJۥ_フHbF! Fv8sZz턢'eQAӵkaY5h?l\h1~GBHbCaLr)XxſmO۴ɵqz cP)4^vzH<t?8;?~:hQ(ZgTq\|܎-|a,-Ae3 53+RaQb ?Q[p#ƶ\Q [%I*)SP8x 8W~a(MÌw=  UX2RE@f!.0Mg-y'Wk&4B$g̊Z?0,:(rLvaC}ǞqǍc,0,KE᠛/Ake@g#{ l^C)bJñy|[ZX˕`$UkD>@e\ϭR?^he†h8SIbaJV\AR%kl-Xh2gΪZ]B ? ct ]3"FmȶQm'a0!K F@JG 0)m&Kp7p.p:fgЀzJ Ϻ|4z(^s__z*4ӷ$:O[W*vl{]v bHi&3 N)U:I30U Yk!BFAeRPo :FISL`+d2*`Y=gqk S# ^kp{.&OvzI\8]i1UV7ӫ[bRvtC?*`yNO(/G7}|vAɊ3wp>" L-SŐp4dA AşO7];uFgOQ{s6 p&Yu_# ~-&[9= <͍i;Tkr]4bFj)_>иfA1LDb+VTU+)%7O޵qЗCۯ}8D G.lj&{\p8<8[ڝUupd=YFe/ryb?~pOabbI.[| ٽ_?B|G@@٫t/UCbg!U,})nV_:[vK$3'4$-tVt3[\i%:>K)`pY&~!(/Ow'ygIvQy㿛uYUw3n|/A~·0EÏ^J `MP*hW_fd :۸}S%Lu6-f_c4g1 d*-ѬDynB^8(ҘIv^rMi!Vwh\)~n7Hl.T{>udY Ho~ Bi˰!(۝yo'iS:Dbb3$R>ˀuIC7}H.GlD0tyݡܧj.ѓ ogvǠyJy|Xa{}n7 v ҷZ}9߼mҫCi|B+6ɽ/i3_s 22e%y!ל7}ߦ]# 铍=GlaN9s̅~Ss<ĂhQhIn{8~vy3iЂ2~7wU ,dx!,1v^)ة<ʑLG= >:SEzӅYi]P:(R^0UMڧU Ͼ.oo_ǽ ۅf@ia;T.wصG;FY:,H#]rʃ̱[_ʵ.x:"UQetNk#C1?6g!D05<LrCQ qJE  `\ݗnA0_kov/ F-7_zvꅍT.`GLKL]uAT z}c:!(B:Zz. PG #'f9(_pIt_;Qn/) 皐 E3J$qZܡ0,Pa8iLh(a`X.4C-S:頳J`eµH9н%:م~,ج\S-'+ Π@)݁zg 0HPqeAur.Z2eSZ_& j欪/T%hiU2~)$K t'*``:qu Ea)*]tz!Q:X`A}804VX;(m0Vpd,USV::PKhhYF aE + X$RxnaE df9V -$Fc#x4)zVibaZצfK"Fe#*A9JUFx0D( ,wUV1 j _1El!GA& $qߵ/>Ok5 CJ1ehꪹ.'dN39*?AIQ^E%d),Mw ;:H`=FnUy")%c?ĈmAHtq,  lrARՉ-H ѶLl}Hl#@(P\ ۆ1YjA/2l PH} PQ:|d| aZ ł=XnW씮_ŊZNNw@aT뜧29O-PPsOnVAWK0jE@NcvoKSd0*(SXCǀ[VVT2H i -h}G$(U?[9~G5,79 2?~ەyo3GO?/oLIߐ-.g;ah׃7$7!S&YkBVr6bL>riDbTlFO %ZP]/&E %z*Cn@Ds^M0܁HUem`H&=ց6@Nʔ*G4Jwxun^Y+-8&QM\)Y?+m6r` p VL]M3Tȴ$(}~ ou$ O!*J.bWFgs[ %(fL(Cq~&޲ |iɥJcf.fs1@QDPEgljs 4XťVeZDRK!%Q1ZRARӾ8wg^G]9mtS38N]{Hv m<ɛGRk$pN~Z.VpVvmo {s].7Yf/(%ZM/vc6]mi̺R8~,XD}Ss`> Cx5?v?6yx} j10fO÷~cGb4Sh'O-}v (ƨ}8p$h=1 12>*O?IӅER.5chɔyUg tp6*LxR Ӟ#A҂** '[(eDWb\zе34vd5򃹐`ha!"0YH|::4R %D{O 2̤w=NPB\Jkʕ:1j!+F!ת%oeku]DVjZ'y+ "3 {>p;ṕCs&bѵ<xI3͆YcM\Ϸ觥ms럗A*4K&5;w꜡+O-c(Aʻ;Q EOMR9kܭQ$tw`b椩R7UCXnS4%?6yLyWƝ0pJ]}Z4 SW)] SSe ׫8`)ӀQ6 RnD8֤ ؘh~cCEg{Tj "aL)_!)-!>ilDԿ2i-/y+YƤ!7rPoM.M9zzblw;fldO?ftXiF8<5Gs,}it>;|o2+¦O' ׆Ε@tAi]m1YJ8m~~:kHAK܊N==:S C|LaW79>a3PN&M,7xU7h|bk&M#|Z޻>kҌ :(]r]@{nBS558Bhؕ#_2qlo7^Ӽ7I:.d"#е(ކA&f7:Ch>_jޛ?ȔHjӽP]n҃CYI ֽ5K/CX=s=eP hM"ɠ+a@u%[(ExEK֕& 2Cũ\":=tJ*MeOFhZ[W8KwW:X?]oCZQGE⥢s9.`jM%092C3E T$)\{ #=u AihXg|֩?gfT)SydZV՗پld"/ެ@Mn?$r~pPMmjM}g POw6)MUܔ?B N9n} I0|J>1Qm֩>>ʌ ʊ,vwܳc"?m"V=qs _̖ JX.ikϵaG;Z a/ʹ?ԧMUE _$(9(CT+?bZܶšK1usW[vŵwd3G'5(zHYeσ"ntC3Pe;o-f&Yxګ̑}hgF޼{ 磻c^3"`!ȫP,%EnSTp<i((- NHjt ,I؇jz6ƫ-a=LWZ"HPƈJAיh d`T6vB/ ABz*ګVf¤s;p`g@$EN>$Yk("FM{ (u<>P%'rjl0]_*Xt.`(`]FZ9ۋF譆g\}Rrst=VǷ+!V;ޓ%pu=Kژ.IΒn>IB(Pb+W0rV`!lP)WP&yBY#h7u "U8 ^!AJ /]A c!2ޭ 8:Y @qҎUʳB 3l2Y`8vv.Vel_DT% ҩjfT#o) ݛ>: 9ۛ"#kQeRXQɥH&?/Vٞ721t5 hOQkiT$ūZd~jgJ1;@\Htt{VU]{.'Q*b=. bF¶Vhh_jwqPݢ1/P#c\*~NqkU.hnP=גzzWGZJ44<cm)oV;]bRve hU׽^Q$èZQj~{PzUxqhn QZ^BpN_C;=(EN>_6ӽ>92( PQ&O{jW>ǩ~myOh =V"/"5|k驪{IO.i2)jĦ^$&Xԟ]׌j6ZQ J<~H8umCPJ4 MԎZK[Bk޴5z2Z4(Y:ן YxX?ڰZK5urqFL+ߕ6WTW{d<-N̦f1X d62\aЊϜ@[T{`kF>(n⎰n5Q|1PaBs.[kqv"@˷#M8S[ʉ@h.0ZRfL0ΘYLcB$3B{m.iA!rz7Ҕ+lS[b[G'5.Bb!!=`DȁʤXAbE6'&7;_>,V30b\gq~|z-N)sO\OFf8"g }OWWQ6k!{|x`FoM̃/[7sn= ;Nx1?TTX)jM%5,K4IF2aeTq`"Vk0ST+wܷ`K' K.t$V QęЂ$4b \h#שN .6a5;WjRK,Ʉkg ,zC\%b.Xj.(`(gDb5) y9Ps@$9ߏpa CfTYJqSêY"-t!rŒ0]#5t6ofJ?atwք{~0 u L5b< 8JH$!/I.[k"mQv_=ЪcN>P̻0r.+ͱ(5F9,=MƩR·0Bώ{%]'Fo{H:'e@lEX\k.l]}jƗh3F3> "߻?OfX>lo4!)xY"hQ 6)h@Q,lTƱU*1)2`Œ5UO?A=7j= aЬn }`YWHtFU.=c]Ƚ Pxӟ WDRB :|>uWQ!0/_HOm{OXDݕq{ zH5ZʯVQ6Wq&~- @I^.ҋwvd[A#Sk@ J i9 $H`dg۫aڕXCuncW|=O?\~~t̟ntdE2@0"CEX|*_B-6L=Ht ɣ~IAhMk)56$b>H )6qVEؘ]:~6:. "5GLQ +@fzJ ˅PR)U&LaJy35qk/q)1X NR R$JRQ_ٝE4u|ֳU^˾eak 0{ {F2 )fc'cg =啡bW2B?gooj`ܮͲ087G*˷%$s :>8K{]P:e)N1]98RT!NQT1N {NX֜:p ;Ne9-W˚qY(%960 mu&?qŎZ?5obr&B`7zWtn1M]KvwcjU%߲mG)*JV6 պ)YJI2tV\[qaX~+٭8V3)Ca:Q s}0 nfVL9)9gSzZ!g$G䬞;W8HvB?Y3{=8>{; gegh~5tgHoaay7~|īo 7Kx2WgY> B.[DȷGAz5F*Un@fZ||"K ##2PD#owaLy7uࣸx砨h,,_}MygCOBͅTr}!nًnTel6-⭷5n/,G`x`zX k|.1syXR?rO%.i#P-ܪޘ@K1싮"thWy:$™0IK2Rd&Չ`Hfskrk* ԯUzvpYs>}^(*VZ*qJ\yEFc-bVR‰D4L-< a}=|Z{Xm4in]PB^!T *dp?Ys;]J!An`?=]S瓢 H"ӕB!BS_9Bw%B1pH(5[Y+DwT>k8f A 2ZXK3yA|dZf"δ6ŸROm[lSM{^i%2rN(Ѫ~4q_Lfөn,R;Od|HW[x#f+`8)ؼhqju?cx^NI-ED't2֭=IĭY`BZ:$ ~30޳nTz:n<(":cnHLܦ[0E[hLi>%ɸ>4ED't2֭EL$ E[nnuH.;2E "%U>n>*2Yx9%g֞AX dA@a0}Ӧ3Kp_ Rn%XHeH< JALr "H: jvC3]j&IK޻]Z"Iz[8-z|?-ݿq@w}3|?WmWָ]4̦׮0(/kW\yE]FzNlYK4- L -p`}I39D)Ú +Hm?=܎d.?s]1̛X˜TjY] '`p'ЂU{PtP*Fٜ0zC˜nLc,@L6 1+E}4~]on7Ey|aA{6Sr3SDbĈ[CM7'd@̵ 3-:FHY3' ' 6%0T?%o 7_AvЁnCl` ȟ[{-?Jn@M:Ux.•y7v98-Z|Zd wh`Us$9О$][.Rm?؅1,[3jc&9 _gY6+*}y?TKy}$.'!r ^#R$A$H1N==wW9U=5_+^3дOay1E8?xC&1x[nϣxcۚ7}_">ᛏlq`ݦoG0?ooa?mzp2߁Wi>+1 \*&Czcwzo:"!gR| f׶IJESxs@2aO 3q>iSƿPo qVly[h&*-ADXfۙs)DmoF#YAoizy;;d2d15M.ٷ'Yp1 [` w jkAgwCoCg۠uNz[ 8w6xޝvn7UaHXGuR XMr2; i"'Fqax,TˁI2GcXFLP=ydBpSdò2D_ZEVH^5ZB%8hABؚ΋8DMc:т/:h Z_j4&KHoSW/X%x).Sy@1fy@^1R2~ _0%*fRS۽AKdW 2m@wII0h%áW~̮lFK!8oHD%6lA(CKJIVE6ta)E:5]ĵ40hĢ~B| Ñ=i脎.%8'?yV~EHGJR)3A5n?! &TߐS_TUbcgGkGKIrjE@u}MlB<  !%F(KUSjחlEb [rTۧ#3mfFW~D1('!4礊/Wh@3(M!x4Yh㘄i H!jC pM1t->m=U҉/G;Gsvoy2Ig@%m._UTbC/RMR _ -^ODK=k-+*qb۞^ F1pL`N ``#yLM$|kۄ\2΢m:PYd;S% hwДէF>_&` Rq_#`"0u0UTcuVR^toȽgH ~q_{LrllLJ2E|5Meƈ>=ҏ|Fˢ8H}^>t~`ũT \[HR3͸Bp)k &t g>(ƞd3&2ȧ%"F+T8)dXPıAFPރ0Ɉu%xXpP5cUG}yvҀ(FH{0(H{O$C maZ^# T|͕TvBJ3 8 }hCXt#fҩY$t p^y :c&R ȫaFMpJPm\wnk+?ef8Q-Vkx6W%>˪,-!raO kj9_-^_>weذ+u^;JooU*yO ~%\A̓t{$['Iޱj7LJobwK} 6^NL< di;'Sv :A*ZZ+f`ʮfxˮhУj$ 6`raŰģɝ']MbP>h. $5/\dI/K3U޺?/ tWm g:FHKu_8`M-9ä7iI`_ 57;,}>s;?C8˷*M 氩۳h@$p 3ô0np }b MeoD4 ]XuQn|dGm,81􋥈uRF6ؾ'Zl8ˡy9#|u<<_@4 Kn!v|(9CmCgo^; *cm݇| UO?#[ENKX݌t< }JëjZH\^jFB'7- ~6i Li0J&wVT 6L)2@gy# x 3bBEcާI' zՍMF6)*OWWeVNYVW4E0]ë֌V5-tlGkI:MLHK5&ПoM{ ҨԀ THf{{P`<3Ej>E |zY=rGOҥ-TP=Ml 2dT(5HIM30`Bd.r;;{=j_-J.c[߱wt;6YwÓYJY ʹ>Q;pamaŕ"Axu7U: 0FmoF#ɐ*7"duomzj䙿xsXPIݝ!Us8sicgOI".gnC}JocYZIEjӐk'`ŒB=N4^I۰I~EH`jp`VlmϓUQsH AiU\;P΅E]*=:Bc$HE%ӹVvR3Pwo|ӛ;l}𧺅rV~{T9oK(q[XIɆفrвΕKM/T'|9?]y``eo`W,lb?cl4L<[iwalscֲyhWYgѦj jWII }\ͬVƘݵ{ B=^΄NC3 :V=SkJA Jܥ476bC_Z1 Վ:9:32MU@p vN;`Dv1ZEp9䚸Wv6v1B):Sv_^pj6@`E;D*Vv,&D:J.9"Xb'm8Mqni6>b͑+l?t4+$ML*).#}PO >`CkM e8':(let`Si.,,ͭ^]!N"z 7:-:Լ)s$d+J[S >SV #ͩNl4nD T4+jrOpƸf邊$.F40n2V8H2D`xW!wrC{81ׅkUGϋwXUf榁H⧝>_u*;ͱRq:ؕ9axMdM/T B,;^/%U]2 )-+(*Ezy8{ږ$;DIbSD NxBKshÔP*MKZh4 48^]\=F ~B2)}ዠB *겯rZܷI]xE9[۞\#UDo'C,(kMt<"{lR8j;k7]`.SJrh=r[;op0`wj9vp3sR1P~-e每&qƪ#E]^z>s$ݏg@,c0q #ȣ/ Yr耒"Ci{2r푡vD&ldV+QΑAJM[ WTnn,[޽ #;Uꩱn$`07^!A2wwpX9W MW ͱn4wrxt:tvwAAkGsP , ٠Gl4⚽pcgr5^#ECKneDw4ؖ .4׉ޓc{uAEPu !κ< Qpڧc<:93FDۮ-l2H.o%5L0AIĻKLfpxMdÞR)ۙ\z@628N>"Kӎ9isI(WMI$ `U *5qt9XށvXFf|351⢎brL<{|hmt>UO)}"A ~N⒟8i{@l$w1ɳ٥MVAo`olc_!tR?Vnܳq'BOOL"o" ٳllUյr0۶ȊsU=J?o'5lG/!4bz1qx8T N jdV_ Lqhdv1FH˥&K2Saĺq}8LjӾ5c"#R 5R3I1ڣ3:kp5qqZ@+dV%Y !p*U*Bm-8zV8 UP}53gcxoݱntAą" l@)WH?q>7wjy00Ga;៹o|71 x ޮ[,S V (}_]z ?-_ "~X. 坺,py7tl&xT<}Z^df(_^9-a'}#~2l<$1F,шI(N T8.;؞O]C[ ]1i?ڰiIwߧ /F Qz2:_`Gzz9f \6 Kfd`R$#q ~GF#L"mK S̘\g\ffOӄ៝ k_ Wi,[8E'[)W۰"Á~~Qz挐/Ox$75#8JpKc`<x4Of7 y{>KzN6gLbP 3=׼b8 D_KG5 \qSO7P.d+yQHJΈD\(1ӑRFN8@gx8]6)_5@G W./ǁU~gޛ iom+5tmaeAX^ʛ+xP~V/(ä_Ӯ+ˌp3RJ+2F~ٵUSU iLRQ$c "Vq(Q=ʎMh9?  5Հ i6* p9QfɜpRkXB_~%Yq}ptВlP@ hu)o MJ#P uWrĸ4f֪hV]U_u|?@")bm$Or.OցcHf>2_+Xwl}-s֐WcReCA1Fܔ {},JOpQj'z*LA'U5Y_ .H2Ks''1ZAtYk8qK,eB)J~ tl. <޽gT#Dzux;c 1*v]lɿ>9`usk࣐bNZYh{5p0QWHb7%azt2pu9 T0Nz/Kq3_UrY,ڥ&$bdv*u0t6Dž̇/:Ap^ ?kڗ ^WfNݝktؤށ79(X}0Ǜ~77lMfY_޾=g{lLa_olyar|&<: |ς޷#"5"1Bu~$." i|t+STܿ֔˝Dfr۳s[Oijc[v[Xxb83"<^h 7+}͑DB_9V]kRXX>g0M~rOnKVPKr"}6)̹we~x4uP3xk#gav;,Y]-?챶.1.,-W P&~G p?.bv~sO9f`swmԷ}{1Z~ZqHqQ(#klFލzUG022dm6yfLg2cH d#* mSSANIR$Î##Jn{3ҘSE$)Yj((xN[H k֣TⰑN#(CH3FHbl(%8$uJ\), IqXy V+0HQ6YWR"9~Bmk)+!|ptMSs*꿫~Ao}.M{jق\2։tWoNĈH݊Q ڝ@+E KcB;m\/d=-plCT̨ܗ+\/-?bGZD4ҾxLK>$珳KFw\2Hޡ%heLfBi;*2]Y8lXǂOapKb$ilnuY*9Ww.qtQ]sL$L]&SĕRwyYmz=X[ޭiЋScgB>DW6N`SQ% :*&D30gL>C'1~yywZtNh_|4vnm* ˊ{puurYwoAb|?}/τB K sdضK$;KP’¥X,iT.)BgY"1ŵjY) Q–u[>^-3f:AFtQS058S˺dmI+c)[{Wu4/uÞnxC.]oM]a:pĘB^ʙ;4Ff;51E/.2ޏ|AnN 2vB~–q5:^C|Țpk ^4i=b6%'F+q$W2b{\HC/a\t1ժ^ X Z\lcm6m րIvu!Lz|ڬUW[vrRɭE3lx»c I}0}0W{b9`娄S $=EFI I/N`4"()iJt(fYf43GJM^Ђ=M B! TpJʜ & JLHɞDl,SO1,ILd3&Vq65Ah$e ,$XP34M k֊LڐiS3-o:d~gVZz.&wvSj;Zm%/&| B/laԯTkZh&F*2)SA<} 8`1g'RlS2lbp 8q-m'Rfkt_.qi{/Z Lr"N0k "s%Rx (Sprz6EMqK&˟z*7'DlZO-n)/[uS{Su$cmGEȗSO永F헹 $J%e;ʢdۃ:eUOUHVu ҦI փ5#@1ݞR=X{vu%k{bͻ/yhC$gewc$9 T<ʁR'A|+uҘjS $ZR9WdT3I4cXRX&9 jT!&+7dnn}y'8/e7L?eyӉy/7Qܫm_פ3ߺq qE+9|G>Z7DZ{"Vk݅3l!,p}>i+3yȦeI57K)|,E{6?落 H\(.Q)$}:fskg^=OUK)reN)}b2;"&_FmKj=?i4+6WS*1'ʿ܌ssl1^xJ?^[fߛr.W߲[u˅\z׿;;@9KIBsm$Slg2^.0X3+~U lZpߔ!ֳu#5^_n@hwx`%K3Z0G(-P JtB+!"&6L=Bj#3L4CGhpA܅ hgDx-zȰ qXC 4BR3%8MD2X*xc$f^+d{oaY{BTJx~.L}z@Lꟙk"s?߇A$Zmn|%z\V@`0jYZ$L?ݎ-x訙Lj!vtrMιV0YT( ⾟o@- Rl09j C3@.Pg!*Daqdi* Dיs"2+.B V0 XJ8XsEHDeYmcdћO^qŪkm ߴ}=w65h2/x01%S6|旋/Qn/v9׆ć݁7 ^!Oz]=Tk0V9RA~ÉSDzw*9Ԉ W1JJ䧗|Φxe^&|ay %TXO`048IRwFo3z[uF/Г(!PEs. *cRH$Ò m5ynI!8ʰ*iQl_J|п_ ,V'i:U( 6 -zân ƃoH u6 \*t*%,$x V#C ,[r %vV(_rnr&ARbȴQ۫X-3F7FV/aƎJc珞b:?(ՋS2A|xqo-5nb6^j;l;}ׇ ]k,gLT[yiolU~ݎA;u .>;}~[>9OϮ %xRS<ؔ:_QU 9Y~/|x DET n5;y\R)#5!(&h3;ձqu꽷WW~vMN@N#hJPF%\541>F̢*5_=z=,gnia-h[^K]W[ʜKn\St08`Tnr&Wa߉z,pnYOd>U N:18xqH9n.]% k]%w,~t}d9g/1]yݟ:yb s_2V[OКDz}8=HLd!=HzJF:Ҩ%QysFQ0͓̉~}o;u[e6RZs[#_pD(~")J咫}f1~-JYI?W;F;:_Zҿ6UR~O>I=fN"߽/vdYӲ t̉R͒ s6`Ñhilh e,%}!p<vŖ?@!cVhm#৾?Vw=s?ՎWE߭;f-O~W7%gGj#=.2]s?wb0y4~ ƒFY+" :uiK52H`Ug7"nʾCȸ"[@ 3 a5 z5+JAwֻs{G,3yGng@*zc`ڃ8mGnr֥W!T8X)FjcǨ ``m  '@EnZ7iPxQ_`QI(MQzu~բ)8Qu1%8#3k̘0s E~NT)*Q2zӁUo:.)(Ď/%!CPaL(7 f m(+}ޔhBfh5\ 7F٥{7&_.M97в<Ojk3kB;FMj6ȁ/޽!tC'qAEw/fl>J2-L`paDcK_6_ `_%Tfgޕ9RO:L <g2U!A"o&4264tc^&pd'>i(%tv/$n'u^}wX&SZ;AHU_|87ϐ<Ε9>ar:O1$PKٕ<vg1H@:dKxPq:)~vBJuw 8q}(R`}eG"*ǎs6]Mfg59;-$m)9̉F-]R97M Nb#:5@`z}tOJң{*ڋ8HUtk/KR{KmUj^ ZqɨN0҄mqR\ 1\Hm2{'Vk/MjtrU*ޕb~"Ģ\@ a 95Y22c,2ZksXZN@k KQxE-~J/2v .NL*XDaՕ}MSJﻦ@k*T4LaJ PB5UawJ)rC9 WJ'2S]if'9.3}tb!?O%Onџ3meSիCއ(4d;7G|%򾡼fL'㯪]'x9u&Wk;$(]bɀ-òנY7hJrRȹN],-)ܮp͋~."09`\($}&leziʌqי@v[Ѱ*8P9R*[7SJͬ3f,SەߗGi5 L,GT"dr ʤs28-(yNsQuخВ/ϓM Fevd/ /bbF;@:O(*+LE *[(q4 2Bh $p0Xb7* *pgjGL*?: +%bu!(4|ݕ곌xxMRLFB^Ův2q+RkU<]zj%^5 X6it@1jCvJ_#Zi)^XDhUTݞFhM+e/"/ۗ_y(+N )MZ">V"Vh wZ=m37Vn-YL 7GGn)E˕Q_/,_g7UI;?RqLp̼UdCwVVxaYb|%ҰBnx(G%(j^1ípκ6JneFGQv(`#J ȸ=PTpFh民6jXL_/n&H[JpQt(^ո^ՖWu]{6uu{ l1鶹[VƽަWަ=$ jmh7ԗy55*2 /P^ڬe7a%64AcrÏߞy}%HZ.B"ϭ2Z#OpM{m~O'Wgg׬'~_>A-`J让Նf{igG&@Gl}@O6F;.d:7 d5aKgVDKE2+4Sp͓,(ӕkkaMFHo8f [jQ* KKStӚkV21ǥFԳJ=Jf#{Ǩ,AaCFQRF8cb1@UEYp7e.Bv.Ybn/;*޻ ~IJKÏG}M3a軷׿?}|s0)gگdvA<<]3T9MS9_8-s[ ňD~LC(p0.cxL8vBBd7*5+Z|1J5*Au?_/x7r+T] {.]LWb n5 )Olk*A, f#C4chXx˸fZ:[cg[ɀPc|nF cĀ;JKSBN#,:+ LP#Yk0 pFFV[Ke ޜ{nzKߣ ~pwoonJ+vW'3jWWogЍfŧzh)#xZˋO+0xh)#x䫎|%VR:rSkCˬB;rS/=c{uԋlό5?^GGt.ff<`&%1[`9Nb=c 4d\*'J(8`򲠞z4l@Y^e!^ {?k Gu5^Ϛ ڈRK'ADXDR _C?H/qO`*TpfH#Ұ}i*-j _`XQ+y]*0n(P\K^"X>lY+TM@h}._hwseREqqutjUž[&xkq ~3UgJWNi8]e.#׼@O>^^>.@Z}JH´Ib $IKGNtR*zט_TNdCc؋(/>Y&uvQ"Czx({ҮH:EiWRmכE$JKO$iURlLHV%Eѐ(-!nWRCJ 7d4 ID Ij% NqrIZ6{9-z^H>DҢ%AzngV]qCT҇ >Y|+|埋Uo9*qT K`;W<^"<*c*@_wb{k*HR)@i /f#ye"-#2(TJ74XұE?5iOnKNX\穕4|xwi#ϓwhDj00L/C6Z~"tuN>@&[4KP;#˭[W߻M;2Y2H\ |Sw5GP)uq||()(/:pEڔ):XF@0㩢;U<WaIMZߩ~ɇe# '=<6} ˯!b8T>lq<̊SX7 ݜbՊYZEzQTGUZee!9hl..F+FE;8J iKfer7&rluҸϊmanzN\\L5!C٫h` JAbdVJ(z .iĎ1F;0pda;śs^﻾ce6HnWl0΀PZqCx\!\")Nf#U8P6H夊x8%kڕySobCI$%Pޒ,`V.%K! xOs-b%YKy_au ,vδ3->eųųC0FIOV.zϱO,%h_Dвm K9g r7`l!J?ukYPN'\S+Bλ@8zuBWMT_R$2|>)iB|㉐Q=U)$ڲw•NFTe_` M}^|7Je0)%z, 7^X'Kˈ-1P SHIaBJʄʗ~vD*9XzZves5(I5-,hʭ.0/ߴ '|T_ \D)˂1)M<+h%TLAPH$ACYiS4HeV`ȕ!gyq(r~P׾Q op}oP=jV`^J lA,x,8&y +|Ш' VZH>Pu#޵+#$ؒ_r(n .>phVK*SQn~4*7 ?&aF rQJF>+oe@JD9T0Ṭ#4)h~Zb#!*WFv.%ab#𞋒ҳ`R'q:QYzP0T,wo;XHHoaӀ;x}p)iqɧoP 0XOW''yo{'߼=9}-SQhO=0?b_irPq09_m?HQTC2f fP&<ƪ[1 [R ʣS+Z*UG. BP5ຕے3iֆ-P`-+knH'+b`x`i]K!Y4D$v0G9; ET #4*ʫ*v]#8~s(%Yj+(x7ie\~1b,y{]KE6^ \NJҭaLm/])v=W`#Ty|VaIbmdG rNY.X R%F>A zy1 0%'CxDHs X %Lb˔~d[IO\d<"N+-Q3L=B` !)8h w"k}@&(~P@2!¬EBްB8oӘcL#yAD򄀘|=R`Pƍ%}KFsRd 0؆.=ˏ MWVB(RT}E8H~8 &}|ȼuHSpO671əxz"\]WN4;KNJe=6k$(V7%$yL%]T/3?-שG@:U$\RWɭ2J$jw=a*J |NW9Gwi&~>|`7K ߾^Q/ҐC!/lw=O &:O>v=KH-1{ieş|v>\_]a"/_+R| n2|",!zX,j+;_ߕq"5z`P)<:KD,<&+\%H|Ԋ8K{|e5[1NIQ1aB3:&~fS eVܺCJ{ >϶3,HLqb9 :.2ۧf `-Z/p#**ƅ hJGilv[n:ư nj!Dt6Kh%T)RxqZIXӂwm"bvIBzӕc̈́>~>v=aڮYҐVѕcSIn=Tba ϏJ\<ׄuġYrˆX.Mk&BDoNɍ,0(ɠ ^kԲ}3T~"ާP-ư<hq̾&k ?[%27y~'>H:!}&+>~-C M/l6 `v(n_+g* (Zҟ*|$E@S,;zurwSd1g~yo3theetq]A; jvSY!Wvǁ^ A3a^m**FNY;0$Ӻǵ0GrdWs:m8Ò!)!б_ 4u0!Bdч2LRL&;GbpaSHT9$}mGuj@PZ~ 5tYGRz#(7WM1Fvv[iQ-_5$U;ҭsoB1[bm7sS&j ați3X׻ws_C4x&%:)|WYrQ".b"!xSI22Jr0 œ("1jSrTVmgr]7cNxE]{b >7 @6*~~~(͉"TS@6c0eν 1qi W.(X;>Z>53R:}k |qڋBm hr)4-]/Կw^,lZ7h{ީ5\"XsiV lGH{4l<PU #N{|de ;CUp21Pk1Q  Vyp-#8`cJkӝ{Hf"z /i]r.i-MlÂgx,~.`jMDERfC%R[٬!~9S!Z/f]e>߫&PTD,35;QOga=,/>gp)b|Ѭev4, di8gK ˜ ٵ/~)W|`0W B"hFUV 2a{W79LlqΐK6y SkYsoAD؃[qȪ:/ ^_g=|6vK=1!uf ۫Znӽq4h.kjq;ʭz-h<{_QdoSiE#R-ʽN3˧{giq̈l*p♇_I*H5a-sGOs@1˼^8vEp @q߮f=GGknmߜA$aIޣ{0"q>Z%㺫 T,GpR3roqԁ 0@\zFB@[+w`ZhCip4JO҆"mU^p6ZbUå±Q q}y6×r۩XhڢƻC P5cDo#Bh*@T )_;DS QbO9)a!) ?"JAHE>0p2r@ŵKts$2{eF)7eڬ#eiF>k~Meάs0k.noc|zh *i* R);h_dy5Vqx:v'UN`scC  ϕbXCpjwzܑ񞁝A7b\ʀ+PQ86NW|Q@hLytrj爋HS_u98 C V.">̂ $d5BG"%fiɘ:[9Hc1xNKiXE`i֎ UdN@#n8@()M84z % ͺYߙu=KO?:,OϷJuk0e <gd6B,䎅.Ʒ"|'EQ_8~M~WwpWV'^(ůB ?o/x:~ VfEւۥ<#c1 Ǝ DϼP\*\R=gתvZzsNy=kxNTf NǟAskӉK? *ڗx~wyj iGTг덍c,gN/0[|l~V<=p-IA) L.q;M7a^vYdҼ8up FAi3 Uಓ%.Cx ` OS|muͮ3*sgCFB ಮݦQ|ǹ?6@2[}_P+ 捥Yd|S"|v^ȗsI/WTBٴ;V)m R:*\Jג,yZFؠC#fcECY{<2H& >'}xA &_i5#>Ëo^0 @=װ66֗o6o齆Y6>Өjw> vIeDag%)>>P\0-Xd<) 54fOSYn^v1yA}e޾L:ˤDnz&ҵSuvygm&+:iwwBX^~(Fk]uo*TN5xE]FLnPJ 5$"5ZٺȪ+DtʑMC eS=-2Q{$P+UqCX7zVDv3R{|c5_,N#V# <8ڪt?x]JwM$K1RCJ *;'f^Չ\^Wu4ӥuhŴ \w%oiG+Ta5l#B&w-gynPuܠ9CsT?1u` # P}4Co5 WnYwv+3Sm̸f'7zڊ0t7'Q-(\^_Ayр/f6ۑIx|MM6zmo17YB7ٽ!ISg5mɾvIuÞ3ìӅ|$@Y/;z 9i. -L@ %czv^P@hP7YGC+3fe~g/, wwCU mMՃlU.sB"=廋ꃕS9MhڛD`z=0zJLM{aGkIwԿNOzG`2* :E&HP7"u{M{ГNqPZGR}$ZﮈڳhuWu#:z"E7 {n=ID{- >I^TMϋ'Sd.oM̟BۥZ?ۯ0A{`Pųmj׷ 2׻m'}|Hztgu0Nhgwru)ԢدYhyLi;K*ʾ|IȆ&-#U ̶Ir"TK{ͫ:N'ly)I$Zt5GF[5rhܲq+yT=9~l`ځT.,LMzy*L/D͋'OBZ+s"iNdBvON]ZɓԥG߻aDхyzEcY#JFWO4 ?xj ]^(@XUջa|$q0ꃦF20ǬޗL8 gTK<"gp(^6G\@kTB@h=4f4oVkR6%#{GEɯ[n)S7 E VÞCW(7 ĸePaOu\ [?$/\R*Ȩ@Bbrh=o2/f BKNj7]FDꖣܱr$\ 4`n)p@AcCt=B jlrc㚡?KKވ?7DMuVDLTn?Fu^+? Ϧ4M{ Z;DzB![ \/=o (R`^>&К7!eِpvlj3<x xhS+g[ ڔ*}׋#Pʚj%R ztor0[Bʓ:֮)ndlO{Kp+޷wM)9ٵZ7O/|f"TUԎ+Ƭ<BSU7!%7DH1 ~ λXP~|,QA_Hwys3 B~\L{ a>, &7 30 织LT L:QLMl>__P?/WKs]‡P]պt1^gt+;0?b\&Q86E2kk\]Xt̆eP}߾y71|-\6;cWp.A*-HбR8};$<&pp% Hk2{ U gWWF}0]IW\gb?aoA~0+#Gb}n $KoH! hiƭV^X`ɜ굢S*̺yfD4 d*^Eŕx%ho$iۖ54cΨ3* ]rȪHC*0jx%h!-5@:8N8#A`^* ,"Z Y0k a mV΅568 0_c VH0-I`Q$1JpC#hIu"\ "Eo5` %ӌaV!A&I hgU5C*ͺ52q>~a1sj:;--U"p8_?t\W)>7b}bSF5PN^빃bMH+4WONؖ5(#V9!G S" $Q!BDkJV@xR ^u#20KqJ; 3q$ >4jlSdP +Z Z#=#^yVjC!0Hw& \L !BIz" Ja{ Iw+lP64X),FAsM` Yks1eNFE@NBW KÀ #mi"b2&n$e΄&3 %JoadnuZ님=Vӟ8v-TAH`,arJa4La>ڕWBLsOBE@眗TKAPC 9eb̨;6,jhx+]P!$@(%"hqbRૌ vqA{$L)&IBe !`B+h I'9 HJK K.M`@4Pxy%,a,K*tcy,LG$M,| s %">[o 1 0em",8 lM#-d0zNc,BG ( (BAPF@^n3E!"1<̦g?gbЅ5\zUh}i9h B5呈Kbl0VF4DZpqVAb+E D&^B~]W0A_(b0Hx/ d`J i2109D+TPBinA8= {5kvК|jXy \)&XqSWxpnӭl9y13i̸_CvgٗY1&Y+UŴ$?<~+0i ^(5ӫ!F1Bx7A2ӕOn[< zǝyXb2{ꖱVͷW_ ͤ|q3)Mw(`b}1:8'\.~ԊԂ:–#P Jp%#PD >!/ER[obH4Bj-@m1rƕUmU*ZXD@9u9-:ql RVgċ*!4HCmACo+Sh1 Q{yDN,w!h#WѨ]g\ Q'i-*vyo-T{u~u f/5 SBo\{|7,z(1, t^ 1w 6w03wM%F_u f!Y㙣̭}_*7bܧy]9SWGDf"aP8;rd!!;˙xjɫ0¯+fmPB\p o~ .YfQ*}hJ3[LRk⑅gR(w~qz;0Pa3E>+;ܠ :4Xmē3 7y::ғ}x]$یr꽗`ý VY-U<)}(ćv~ʚyEQq?nAWFkcb{?Hfش 0Aç.%3O?'G>j<LR$(hPTgKY ߽ 8|!T+nU.a`Y.2U zi n|:H?z52C ֒$(B3NDjakq 7{|&a Dnd!Xn`\W`o%]dá~~s#u|/S7* Ic:̧YXB`?C.M'f=$2r8֜SzdOcYAQ) )0(Ռ[ cIZ+xDZLuaD mPWqgYƗjy޳%/W 1\,`V&0P#l 2FW L"2lN#[,[kʉػ?.5mF-ra n?+c]RMʻ$ w!,D8ibxv]kGr5!Sϛ ;]̢Ɉn]v\~eYv(oBryBƻh>%8Hyb݄@ޕ6r$B vb/ܰ6<%)JMRVIRRX)܀,deF|qdFFFSnUyPDtQƺo0%4m*ZԺ!!o\D+ XqnX7?VEDe1wM[ʄnuH5e"T,?P& Zb1K(b_Fʔr60.5qXF>lGxA7'InXۻk36<ѣ8J?f(e%bY.0r U@R:PiBB}_%3r|=?=c''(cr8 [ ad4ᄾ$##I $(o&PζʜMn:y^pJ9x|_8 nB{0ۻyH;ǟx˼b~cӱ|Y߇hF̧/?\i t˝ԴX N1_ -.=Dh˘s$k`7+3y*rI6[7@\{־2;}iHf}† r0c UKĄ>4 4V0@^sT ^GcoF*|]HANҬuF˨ĺ"AD|o+1U*/!lq,W5Z8 2T6KYto/ s9>prZ \HI-5>@Q⊵W`p^ӭ}+(՗䵗"jՃ)ET9Zw5k%]T'rIiAcAٟ-i(yMŧ+Q8}Sc^J]XU |%+2\aBK2jG`)FZp Sf}yDM\!kSֶ5HIiU%pJ95<"~2ݭ:( KelF ̀>Ӂ=R Y9Zdx"+a*#~~)T acMF !]{(`bǍv>ᖢC="DF;{[T$QV@PSh vC4Oҕ+̔VKp8B |'Q0r砏g SH7#%l\333|PgB u(c Te6Fm QPx.Hta2KQcཟ9݈,n>iT&xθl8嘆 6=V`dFi M gj;Y;)2ރD bD/'Ύ&n1+#rҙq ڒaI:7rޮgX~3gUy&Xs Ϊ G`R$1pk5 01r pxH笪gMa9vk$# DRFȤAVagZ/ (,76ƮJR!Sܬ&FJHnh,X.hR2•1( &]ʩVJ`e9XI\ ˔6C!fK`Et#42%5i#|I\y~btg7&u](S]}q*RQկpZREvYu۳5=?[&77;n_0CJ7>᧿N$5cd6qMWfC`:X@gDsUHso9燘C{?B'=*/?wt@"@>~_'OR'E$:o <#C8f'z}d:[.9>ލF[ 'ݻ~O@qLgZݏHj ӵɁf!*X$R(^(X`/TLY\s9Wy;㯇)FvLX0D`mj 7]H 1ut=U"K"u*d=P+1ʔnuҬrQ0/v|]r+4 ' q MhDA_3تLsc MZ`f1E(z6T >sr?MN3FCȔ*X Alۯ#p MZ v O *ʘ23L $#BKjwsBV\&|uf(vΑ4rȾ,Y[Zgʆ;K'&E3h`ןUƦ,oOzfٌ]Ra(89l{| \{C_[75`L䏆Y.v[izibz_{t7>bIX|l1ߋne}S)i1yۥ}-x]bmGTJ[3k>--G:cX'iSKՅAЩ Mgw6p| 7`^ %uཤD8GԤSknB n9LjjKS2̦ $j;m.p,2huH Ј`%ݣq"l.tءBe}[" H +V*W hNP\KBMSbqdD " Kz qPEQQTgĈr%o<%̾ heC@@TPJjd)4|2RRm*EeT(Iʷ \ $ui}g2k\ rqS;q)AҌ/ SIFQs*kYb޳m,Who KC899'm8mqqݥSR$9kYRZHQJEUh c= 'X]8( @=Pkf/p9ς-~l`m 7hB0V$* \i%NC fx0E C>^P!$18=@sѡ>wPۅ4N3- v5-r`4ؘdcfnA?&<wHI1lE`*>>sF):M[<9Z$vGWyMV@Q `̺@hv |]qZ8{2HQlڰl7t}r(I{K`c6dNYG)Q٫ .#𤠅P!^'/ \; #ph10wБ(DTBjduebUWs~wY<|EeþСN* .nުp9y^]_GOǸ;kƘǼp8wj(Wxk(OCχp/(dz1R lkH rfc޵ƛǫՁ;u!ďOQ9Fzq;y*ЃV{7ӿ'+ Hג(qe^1?p#R9~)+V4g- j=1?A5) NMŭ"ٲQE,l]A-!L"ߩpCVD "QDjj?8 WOcj*=[vvas.'&=XMM!YúR)cisjו5kFoD!zYM6ʅdDYr.IMqZ5}\!`+fBF&P6EJ; QD[3A:I:AH(\H tDzCmJQ`.ٌEͤx\BlݻB-uFA%nNm<_L"z_}tppJRw8yN˪,W2 6/3s,ԝ8 ;ݣEvr@O28%G'CpU!XXWA PBun"U ū$o.)D| 43FS##̈"5GFiCBr{yJpBV_|VEl{L5uQ;z^o)4k$IcIt65c31nEޤ5tZVeCS'q 40/CהXZ Aֳp7&YӍSq4j-CQ0{6㵴\~q*.O`+R#jST.)ҕиU˝4 n 9b_g.R)֎;F2y!SC)u))P„$ !(eϠig0nM@##-BiuUuWlk+ٹ|?;9\~vK#(`ٺUƃTi\wcfZ:t CN,G oM_•r8 ǥIu*&FRn96)]%P.0-FAN,9Ǚ$a+y&tE9郭<{trQ;t%C#(hF)UJJ֍2 ;w(u>:ĺu'` lXK SAQ*+%ߞfe̱unt8c TN݂epH9G^NSFhpT06K?j$)2$˻Z%p l H7dGDuRvNۉ{\֘rSWՕan4Y$]9 pIp$y ٯ̔5*'P [;Y[O\=" Cmu][&HJ OKGfGr릲i4mVգ"Tެ'^ۣ&3Sevi0XP̜1AFe/]n8LO|ghMȃ3֓/YS>8+jC[`Q~{省[UՅ:yZkg?bAgyѳ,8m=6~W2HįDZX^E~s^c}+ +:l'Zꄆ/.a\癹X&ߌ'~R/ID27i ʝVhv\2$f];/ ᢛp8/J="ƦX\:nJ#CI9ZXgҪ+#k $W ׫ ҸqE6Ӝ0Ϡ=3>38}0 yo=<&p fF?ԯN_?}ٳW<ū_~ya 2C;8Us;ڿ|8k߾]&^oۺ Es~ne빷I*@dn&<fVx@xi v LRvg}ӻ h ݸ _v]f -VU;]t=Nh04癙b!0koG?_N{X6?HE{58K1>ZV c_Ͱ[f BmOzuA!~ӛ7'?I釧6|8f{Z )|)( -?`9W<;9 GXl_·:[W}Ga8zv{ >srϥ)OG x>Nu WL߼'ֲM(§L[NT _߆|dޟ7l >2psp馵wASg9Ҷ-?$4jʱ6CI[K )M]7h>s>=a >dZXOB}h,@^Q1`W6L9l p5kAF9dۃ%,"..6ɿ,h GDazq@N7AI+=N@NN%&1"dbx0žɔ!#RC5QPԬ96ҚHIpIk@+J8➤Yz$S|%?^iXc{P>sd }@Ôv"CEsvB0_%0dѿ~avf޽fb/OuQ+2(3/ 9/p9+vʼ2ogYvʐtK8N4 ,Zx$ FMS'$T5Cɴ:=H:$oLIWh# #D[(f3} -S3 ʹb/Alvu1i WZ)pjΎ(78/Ǽm˘>A)ӓ6%9`PH8fni;RZ%wcb-yQxl2`GJzilIXq[GXeԢD:ay6&#J}z^`/i/S@4Wh\@5[zK;-e -?T2 ! t@n MJW@x2&:h!gG(+)8ӻd pC61*P[C* >r"4aږy6&KHb9qY!]/7M񉕤 ,B]FUu^ܯxmm9TL*B4Հ#PegJ Y,\0h/ON5M]P 25ob/L:Rx"ME̠x,4lP+(csoXwK墾aS|}j_%S8w.=KuQ"=yZ7V1:nɫPٟgށ9;sx~b׳XQ$G4}^|̤.~0HF!$sBaV1Ɯ6~ G: ARZd&aG/]/ a!HYͭ>T;6vwۆ^zΫXBNBr X#ޢ"i G]wka,A7syU1uMֆk!rzmxdmF⇖m;NIMTշJln =G7У6٫0wzqD9!FcA|~@=! -4ckɮ;=AQz2$:hh;1fM^ɀ Xjܾ]gwg[T$nR]v:[{[-7/ #4xs~Fue}_A)_{~m2Rݧdc/>!^FyUniEIKQYՓmkJ&}SGؔ.Og.w+T6ȯQ^zغn6J>tyba ٽP@o&mSgʅjFOGK&O'2r/ l@RZj \q}*yL\Ggv[|Te跁.ݧ^=Eu#hoMF<m ź}F@6{A@@t:9&Jc3\Vbc{"NB&װSĤbD F-Rt,m~&}A٠:&N0Z:I@Fjۮ߃:Ư]?:aVHT{Y{QJWzFDkY{We)(K9#ع}ZIp8|Nkx/5cY>5csM ޜ_U^e' O9˯0&R 7x2o/~ղNz"jVf(pƋM.Y/Ƌ,Ϝ->՛}oHue9K5sj+UJ5+n<8GêAm.ӵh:c:\$"o^:۸*;$oֳwg?qNHNX yBCusuőǛJ٘(}l2 )&^.͇TGխW?֗/.8M5f(X13c"22C6_]x\<h-&Gk,7~?02h;3 TEN@ TNo48;#d3N:3DBw.VDQnehQ\rl84㟛34"m.>_^ B.(Hz?0s( }ʱIT)l(뺛} Cy 4BgxƛБv _R'X3KmZ9qvjs$=Ѐ+'{) o.V's8N/Q?ib'! *bBRIQy5H6V<,qu_2ҽtHs~VXeUkBP]xH*$zbSLB f d4KP{Fˬ<-/}ޞ-~83<Ҙٗ-SqVuDt]ce0:H#mhC֬^ՉXZ3y`U ]+c0sA}.;ݒUڥ D!Et2P uRl*FJt14Fj)UZDJ"8h,S֪gΝ1Fr@i$/X3RbN0OeqBUƁ<.GFZBVm![E#;^Fj?b {xTbUGbr <.~Uϫ<|/K|G]WW _,vd ΍r;s`]Jبر脨k 'W*.v4.'B.&.32ԋV~242;2P/[+"IL51 ;&RԮNu@8Y%3{ 9Y/I0v|< Kw[t펥%@dLwcYn8vP}vz8.Ke d;;,Iu#tF+Kۖ14HjSlGsdvzkb CaF[:If@C#⇟~@V6cV{!pZM TP]th={de˳: _mM':}YhgG' W8{#j⌋~+ɭӃ_nQ<mG/ǃ7? p{"ˁ*<3'%GcEE[4shJctt3~#5H+) xA=r3"T^;_HRdk}Uv|oQ& [J4d,Pm&WXh#k0ƺ bg2 z1S) o%T =)$)sQ+eTV5@ \?0F"q]"I>FѲY6:x(D`Y$tix.;J8Xg!(e%)L)BQPR&B0ِǁsd"%щ  e\ߙs6 8o-1IҎ Tf,)QWTj;@EkZ ,[͜4:eRЄReWM2HJKYIȊmcZvmBW8ڕeB{IJ{;o3g*2] Rc߇z [%Ici ; <k,irޖ VBrPW,#4wl[.FG,0JlX{ ߚ-c]]N`S:&BFrˏ~`57e7{ڜ6_af!~P$&v&{fjI-;W; z|~ui QgRUl 8w"HЂ؄N!GQACH3>EU&+4MJB*uĂS/[ 1 2b-~$"oxHdɞ"1NfkN⎴+^ɑR-v_JlLܔĉnIsfy =;>WV}x S;W Ǧt 3w7K^F{=xn =wgS4!yhg=8@w6On9H&Ϭ=5M$ 0|8F%#ysC)nTܦ=E_Fp6I4"=*[bX cT,gLU]:Eq ),-x* 84h*j xdC@uA>skHpےOˉXb=&1­;N0^ '7ͨ;z1*L&,j2l$נ=(Ɔ6Wܘ6>?}c}b9-XD}kH?E'y1vlg;Ȳv,r5Tqlg~V ښx-z NC^PbUplrulj-eiɿFRwʌaτJ& CB`mEJ04Ԉk]rT4FõTs`*,AJޙ>_9&*ˠ ['v֢AT#M42MBeJeIʌ<M˄ĸãkt#cXz-&)\VU)q|V4sde̝$z+vT.Sܵ^i,EͥAMew/A<{$)YVwOyMjWv'[^ɂIdBCƚ m)7"b0lۢ-,ܔ1 1{(_Oo.Te}s^LTopln3g;wA QX ED FA1~$&,VJjܤ=;@;˼6/9'6G$ńn{*4VN m0(O׿^GؗБ$Yżhn%8BUs Uqiz@~i AJKh7]$^qH1CVH[@NC%؆$ j#|&X(.DJ[m!'Z#X⦱,n2+TSJ4PidƘ9mu65D.C>?0V݁ HUM,r]GyH;&O!2/S}\|V!sK.*?vn7O庺A(d E*DJkHa!qİ%Lq*)-"T>9RjDȣ(0+1,A#(E D4Z*ο EMƱRaiYܳWZg =S"R(C$ #pb4HRYbXa< Q Xx!"|О혩h!&@uB8Z)6V<֑Q D 04eZ`'4z-~yM6݇T 1 ly@5EĹ6Pb89ס 6-~9MAsI>={DSB,ǐà !@0\@>j0 Y !BB:􄠭CpfqiC]>-F|Pv4JKD p+`"{`Ԕ(,FpGzܺ' SAHҘ#+qU Q<7X ep,HH5o}CASTV--O#1N#%ld:B*ŵVor+Jv˭/!z"wo_s{1"953[b\$4FvdbYBB? k[3z9_JO'v6aʟz2gGC88M/gVPeoU7djAUuܴ{s4}zQqQ}9Q=GQeJ:om!F=~:%Qֲ2A۫m$(>F[IrEaU[;j-aHP֫W*nM]nQ97ȁXà/+%/`NJkeG3Gוgq -:׊TǤ5OO&%oIU!C_&UVIL1w9$}r=[ F H-S;.>U.SRae˟=SMq۲ϛ=Յ~u"Xԥni z}Y}-}.s_bTԅ.eU:URQgbZذ&u1M[ W=,|:T(hֹ[ۉ/P|I*Pnj yˁHP!DJ#巾nv|?v;BE铥L^rPy{Õ'B UĎ}w2wZ<htGDSP>Zp°S{5d&@#4'~wL`CW U|wJ.Oȧ/3-Eܒ/q-'7*3yF7YSV [t)A Ģ.$.E3 @M=#C1U#0*ԟn&o 4M|­3@1usg{1_p;h(2IoN^|=kP}ԉ? c$XňB b0D)Q W%Ƙ`ifϠO PHf`XSpJ>cǩ88#zd%Rn\cɖpygsstn>}BrhhQ#(܇iK".s`|y_$)Un)B9Tu$ STD&"&@*VAsҹ$[] (]IQ?_#c|lJfO(70xč!t'~ VLRxY1҂hܺUك?6X J83\\ESv9,Hi"R$+d9A֝!kʨ$)(Lj xL)v鴃5I)gZnO.@N#ϼ`, 8DxF8eo|6r&L5 i{e3㉍A&ӄ)8SqLA'k foGAÜ,&nmȐ$ 3U`:Hؘp)c%qx9KF_ڍhkZyj7t`A<46 ϓdfܝ\6QȇN/ZT)ӗ0K+ًO\2 g7T={x瞢B 9$_Kbɪo%5Pʁ@ =}%b|u,`}**@"M/0&8`D4&cHӦɄ69^gc>#?$WRRE8ſ;fBJZQumlEN%Uf-qGgSOmǕ\qn;;W(4".qk3ݹL$"Hٵg'Bf_5qljqljqljq'+٪ WG$1UԊ ""Ul04(c]5;PQ#?P5ZC+G>vΧr#gӈ{syʼ7Hl rL{+ M>ΈՈesB|o4 SqSFC5f&q"=^O%qjnI^ەNƋ\2(NLW6U _+Jurrt̡O^iÀ0/ X(wnTwa܅0ٟ/~=xsxvrp?iNm>ߺڸqRJlZǺ=89z MOs=ڠ<&n~0IZ v  jSs7Xĝ1ڛ,e?A n%Շ 4 jȠ(gG>:8٫Sw?O=>;:W9tY w:_~6M퓿<=| c9xuyq|zwUu #;&~==zjQ7`wwNNo(w F݋n90eŸۗGGo.iO\8ji{Niwd.Uwrؙ&tEt-Onܺq)I4لQLtݴcDML zg_M\$+U!XSNF|w#TYhr{ Ȁ$@0lbR7*"n2y`؆7a/:n? <->MkB:>B ݴz ecg$Bfhn;|=mkDNFO\(u?>&'j? !!?};/ 7r~wo H|p^J_R9Kj4ǁ~aš=҃,Lt8sqwFۃ`t5e}?Gfœ|>xiv}~'DA>ݗUK m^[780+ 7dDՓx\,ONYtmg)@/)$ӝED >p0G!_h<C"7ycnp~cWw8Y ߥݱ^|؛n$"m>?1?r5'15QNzMJs'>=]}0%??nl-9;?v޾9Z!9{iwmmY:qʹgUh!1H.A:vM/ Zp!%c&P |elO Nlݪ 7 'v\B|߆9 aN;aN49C390BISeY:"Y8bC:IpF+U8p ! p̌T 3ZY1C2 鐳Xj-5AFAn'!::'oa]8nvYxz+Y?|P{3Q|EǂǠ@-Gܲ޺%M/"uz;dS<׷6(1ސ77(ʖ('$^ja6~p&a~3"Ȉ_,87qy fHkKePA(Cdko]040Wm:NON_#ṕB$ wy5«ܲJ^A))P"cQ~ ]wi[N@jLHdzi7a@5,6L$UHDDq3"lMbXe%54 "19u5!uˑX8 dDgQͰ[Q Cnv ,rQ]-)i!Tn'*8Gu4(AE2VKsni[NM ȋ*P &%^0jA#-DEFb+!;!w[X#NTg鳔$DQgtl9dאn֮]!͡tBtkDr485HYYAT rhȩ[H[NPr Lj|H(jtݖ;vK}pQz%nj]@i;Gś=@5Of e1CF@FƑhBj }s$ +P_K9"'wYμep }!0XnY{K :1U1By^t65!+l$| SE5od-yR1ԁ`1xnjJ\ؔ gByo?w?ǝ/ʛn.]YsNos5]7GaZWTKvG\Ikwކ3s"3lhBJʟ%_nț7i.Slfuf!@4k:`3èO>3"47i;'Zo^OMWe~8%̐`IȞz7mqJ0><5,6׹v}JB[` *F2i+(!%^*^nvݮo:ʔS`[[=fK8TpP/ 3~Ѣ5\.8u2ۿ8- 4ecG4_/:oȓ9R&'T8j>sXczh9ݚ$Dmi5iXh;=dxdoͺyQ ܵh(c0ץ@~>\豜j۪waY|ݗHrFhg$V'xĆB(f_S-(n|_6ʁAlE-k ]C/xbţ70:ިZ F mr{G|_CtKvjkw>kdGY7>OYcDrs\s8&YJfqjm_Um<[ @w5{V؛+viR_.Ke_4#[е{yPQ!8>*nocu:y 3|srya`Wj06!B1qzwJMB RjFɉjW|yœbNC朔^Ml j,\Ѩ9p(+r?Wj=s'9ϳQmMd]J 3tXB٤.h*6amep9'KfL\Ól,l5Aa\t}ҹ{~Wpk4M<QKa٤ߛB '.J%CB |u.H|6EN\+\2Q%[Fc:T(.xI*[P^9?3WJZq'佫,kOt5*]@,]5xxI :'ML„W'Zg܅(cWu؇BJBMݞ4L'UM/ 835?F6/>CSvipUڏweQ_0T-'&ŷd4;LΥh.r4nا}2yBߋSImJpC((ۣ?tz7DR XOX$28;=Crl:Qzb nLxi,ϧ?cpYdjl:SC|2T؊i j 1т.0B!ip ƙACx3HP$V $cC5扂C"`aK,vH'F`'g\!Ze3&XD&N?q h4o%{ :ϣ&3VvdM./7 ۈΚY/kJԝceR0g f&M}܌\s%=Qɳ~%lJ-;%Z+J{<{ݹOh[RڈAXcKb±į0&kSFBp_MҴ94\vztr-GŰ2RC]:۾ 8.~izMV" >Ԧ_vCeQew23gS:M8XQ0՗=ѺlҖ -a7~s{Oo(5|<׏5+:&P1MqDYXqnzfOySQ&122? mlhS|I4<caLv&f;rf ge\M]/^n QE+ OǞ<^5"f)x=2 e -!8*ˇٻHn$W=,GA ГgY`^f!LfKcJꬫ<*+mm$#>2/}̒!iwZCaXMfhy>gU;{40jv75|bNPjJF?pzW&!Z2a*l\5] ]QPʡY(tsA<עmtŎ1^B:N9|5v"U,3 DCD,e Gr-1-4"qh Pc/6A¡t3s9L,̀f Z=l eTXVrR>)Rb+B;5 vF!0) \z5KFT<{#*Dpҽc{DoRy89诱4N/]2A61G:< uz{Ե:ݑֱwLI5}P@jҌV j3P`t}܅w;wl*u΀[.u J"1tb1V:d@DFP^eăB01p}^3\8SLRkiE 2)3eZe~{X]hng`N:~NS oJtFPq &J<%>0Vo"4z\jF5#,~3c4y\~j9Nא[tfjk'{_{*2-P qFҨdOfV쩝Y-jOdOiOa4JOt2-.]#Ҟ6ƲE=5I35{`bZ#l0if`b{ 2t} }=JEuq4 ʇOT.+R+˖Ѡ\uFspNX X 2W"lT"o?%Y @PV8JȘ cg RI_.r VY[FFrw@ʹoT6|xzp:Ԣy*OO^ܤ ߞ/b&(J|˹ ȘXTA!οW7ﺛ vG>;;8(5mj*CL\ǧ?-~ܗX'T_~zrT/Ϸ$ֶ%<%lqR3#JFJ̅j<$]5-3HN'44sB (TŃȤ qENW*p ,3$(!P=aA_̔=̛RbzaPATWk-2J=R8lm?"?:jGfC+:uM˨K%(> q7~9teⵐZRr>eYm<}~-h|^:ܸ6R 猢?_8>J! !oʴr ,VSbPS-dce`/FYXct90=4\ Z? f Ppjo0!tZ&J ,|wq0kp"*~qX2ÏX.,3ti;fbjD'Gv:q@^51'oyѽP|"׺#W͔ï];l&B]e9lHKz19[Oj&+Gk tWr* LڮGmh3] ,}:JߜܾORɡsԦt 4`SW)SX2tkmi)1#`iZc Z2֊ZH.SԲE `+):*O1,,y,DjkUl}JwƐrniT $ݍ W/gQHpK:%]h?>^^UOO.6ZZ0GZi>hxz~nuJHNp-hbj+fM|jWO't_fȏJ+Aű皱7caP0pcF`kn@ȖVK= dAaYJD2YC+~E&kZ[+ ,M~aajQQk$Hs'K {|ڒR8m db%"xA Q$2R`$J2/1Z/.^m^aeduL)\eJB8`Om%P l".fSK`f@˯$tIx@*P !kWXϵLoveֻ@!yf?XNiFQ`G#o72? c5ȆY[.ܬ}nj Auo5ͻWo@]?;I٢Y "֝Y zL)k ‘X)V;nMkl~ly`F1Ӥe,ld67)?]+kiؕw%7fƗƜ|57O`^FnTnjsXw8-:g@Nw={@h#p j`lʥjoPe9ڒCfxO$)NyÊFLl@-4 }BlⰒ93{̤hMKb~r帶wűHXЪNŢ2?)̷Ի4hoxM~YT ",.",.,]Qh!$ hƜ`DqFAG6oI9c<[x"PRtv <)O|4Z䣦1O.:l=W:H!jҖ:HBɢ_$~,Eݢ(0bJIK5q/z+iلe>/ZmBǔ*fMVr%v@&;Kf 2 *`T)sǬޘJ 6M*{i߬8(]%J4_e>VU#a]ï )HyKBC,KA#/h墅V0Z@+K]6K>a^5,̴`Aΰ(Mq 25,a1Wݥ)CsP-t)ڸCH/dad7" qjv"L T=!9'K.K|Ylh >D ٞBlZHT"5D=jbdB6>j#o7{jl%Q%:P<)lV}Ih5܎\׮s䪱#e!@s$R46nye Q<`dÖ777!?* c3B[@ a0E8&bGuir`L~(G(#!TM} nƷy;g^Lm$%w0ޕ0 z(HoL(Nި۱HJ;RƳwKx|3٢0Œ1#'O jFIKXZ+E4Ұ]u{qr&4KÓޞ^4_AK!J3mYkVhKΈJy+[ɽNH[m`2=d::[̓yB+Yք* עZe/a9JYN`FdʌwUE1r5 1к,"-}i, AR[0u<2%VK/DkJΣΕBrW_/tu}ҽYN4c&g|C_~AHS7ҙ4Ȥovd0.e_~EZ^g2z:s"OS]-[Jh$@eo^dV{k׊-ջLۙᆼ'(Cഖ3` Y d3I<ވ (T[Ji\iԪUR+@+WIEJdkH۲L!6w~6SӿǒvxKR5)"G&I!1zirj`͹knc 6(r;U@*Ms43m@ pBR,a0:jgJ!g*XT3FP&J&k7'/ )~gNXsPUfK%%pKjr|{6wMNZ1_:3qcuv>jq9q547=7MsӾv_Qi1_4CȚ'5|CZHF[֤3$i'3:*E{u[=9 ǧ6¦+7MW+$d sR/|{1w90Y?Hػda&;}gY>bpO:C7}yڹ=ُڍ5|<7 .?g\&ctG<{kN$?0o3+uo> /`2u>y%O/'Kd|0 I.V]wh*z?p0bvpqiǻsͯ^DMc!9}/k(؞n'~_z:lr,G+7<1tq՗6ӲKoOx>|ټWO;ɧg,cޫI5p3}2{?gҘq(}Mx?_u5s_ {fɮtzqT : #<;S> o⽯/ )hN_'_>K@aH,G.`kMl g-LB~n /e:i6N/s[&c9D,FNfk&UYMs  3x$y"ў ĨrBӘk1 M RPJ^a0>F\+DXH!;д$r?<",2cc[w+"+Z OىW(mV`coT`#i&hZ h=AZ' #Jŝ'j h&NBBiLSSU&3$p"5^ϱp $nJj?aY5LQ8;:8-x1z j:-,޼g0?\f^g9a .ť"%O3DU{^Cef2"g1m1R G`: P=.U(EF;-gH$ 4O@r!S_8Ę^=Sass ^r",!"O O! 00dda*(%؛[3! }ޒ)rqs d5br a%Qp{q'NSPN}za0ڴ[DZ)IK <%2V{gE9iSWQ##,c!ԉ7CF^ Y*D%K!Rw" YDqR&'js-ȫ:iKu0xX B$+1} ٢!y c|=B"LDC_. m4T#lpDAG+A%ľHA|I醱!eg3JoKbK,՚vU0&*˰BҭC%2䢑eP {a3]J oJK!`)BhȤPbxHV20zSRR" @D 'fJLmr5zH}!"D\Mh̄%կxɋA D{ 6>8<ۆH L0HjK\[[ʠT@!9F\jj X!U5Y'uIJ=ςH rc<"0:߅䎚 P7$pe!ICӼK*TSͶIKw3ؒ9*xsL@0oX~NO9Uz!5{/6* i SxYXd8xqbƒM+tko3kOP|Wr%QS6ɰLL^%"AɆcQ+ˍySv.@~%ŵQ2 ʻX02^RXQ(qnx)Ps95WrcD1vާY H]Qw4ete ybV$_ЩLJɆ.(¥x}rDRH]ȋaKo0R,B>v)VM< je8r_J*ҍ_֧yTЊ]m#paq<-T އP"Spgq>  夐`j!ܯMzn|_8|yөjr4$E )!yKΘb4aԣc)o&(`$;JrZSCL\X "\TSzyyS\I !=Tw$pkV\d V8HRNy0SxgU,1#IYF8 6V:-'!"%jKĚ eJ++WiNظ )x,vD[F`x#j1j-' DDT/آbooVEX85~;yq):\\ڨE|ᥡ7Qz2l&n4v4MgqMnB\vⲛb\6m!vMiJSCZq`Ϧ)XlJEʼt\X@4#&$I}x2F}m7AF# !z;zc+|*תEewq} FND0O΄Lp. )}q`]|0Ën'"`&*qg[wkPuE-XP<}zijw+q-*C5AXlGuefS}ujb)Fp;#tyc3j>,,Y2ط.3dhFۤy0O\X1Ap1{5i Ĝ;RUI@P)+( >뵏B0pM (k9Ff9 ۃrI|-Ə $Igү}w27re~Tgw5/%__lONQ~! =T^j J49B/.CzIdooEd~:1k$X Z`f`{xJ*MHK?ߌ&s, ݜLnI:C˧fM>+ru9{le-=Wۀ3h"\07WOhvh0p@Ovz<|1@:ڂ0AJ+uѿVS@D,_|M`m8'&Mm8U`']ܦ4uQdoT5šSN5J fN5J$(~5JP-F8j/SE+&E0VـK;)G5((ԜSt 0)f1H Ú'Hz,2Am b%͝ŤR STNFF`a)\+po-ӱ*TD kwNwKk& *:Ze@NN>'9I΃Sq$<#!ёm@6Vr 9&kfFH$q.HSܦM6+ėm !ض]VKl[0 0RPn|PCRqSZjE ̅ ~OP{YW:jjH@r.rm&DLBKjK@8$,,.з4RMeBmEW ?v%W"Ķ jѣ=/%4J~ۀa:vW} WUx~A^L*VVWvNgIBodC@7¨w GXESm@aR&"MG@H`Hy''fTQ3`cܼ Q( jCEҨ<2R£}͆H?blGqSV') pʾ+y[ }%`g(c~q2EaQ rcjK핅I?"UӣGWmvtFD`iP"mf@[pıEs]@\:A,2t1&,tDuZ+"ºύ>̙b0%ExҐ]Oo`c TMXҐR!F0wLlP:Q  ٕ>e3ܠBE+sޠؠt-b&(`Xd"B}F 4 X0% *2j .lFTQJHNN=i83?GFzֽh 'yx?,+QJv7`)/q1p}%vӗzj^QnU>nע}i',p<5,XZz"vm5<6@1k1X^غktk^ۑ eNԄ?IpL`Xh3B\B"B@q\7YHmf4!VV5.]^r-N hYLin0j$m6sRK4m`k{n096֝Ow%&ً{p'f5XgBҗ2_0;Aq+Ug!ZV5y7k0? jVu8_,|+t1 MA_Y&g2|nԀ֏jPV6ulEP0ʎCeJd60 m1"v,FMEhIhfE*ߕ Ǝ/ [_ߣLuiSeFt+%B>9n5-'OK.j6?:=kiǚ @l;9>Қ5=Kkzfg鬓!%$@K]jŁVKJ-3 Nk!DդSq݆rr\2h05bbsSC-ak10mVYKu9`ǶᚦkQ@ `N|`hm k¬qaZw-21QJtן -1@@•96wCeSKpꪡJց7+tA$$!f% ~˜CATJ a9d2\xP40XJpq`,*qa`}9IP[:á}CQ8|}4 (pPJν"d%墔RA76Ԍ{vF!F G6+"$TBXs1s̷GvV"X@| Kn!滺J᭢B8ZpYk\~*ANϷ9arN)&x+Aٜ-2WkYo]4*G9a3n~6i=hݞ֩ԏje|6?~;z0GVwMOZApHm|t+̟zۙ~ij4ՂrWF B[=뷏xbvdXV> KZ-]7rG{]RpM\ E"$D@ zZ C EnʰR]!LtX`GgQk87uZ"U)a#&1+a~f5UB3n3VEN*؞W3?HxMҿ{/|6u{NʞIEfLEF^A&3OM3urr䧵$%885GӓlGӃG׻(Sc_w=(ۤ?~ ̑^3];hTÿGjGTYc50<דNXàuՆM)dĵxӬ+ǬST^= ,afxwEGm^ޭ(E{cZ{ŭ3{|x.97So~t49`5,|9WϯS|!yrMv<؉R9|>kVB> t*rtpO+d3 eJH/]oId߮&YyUuٙtѾ8oKXz$7g|'FnIxuTWm̋i]ev8N5r~qջTE"UD5ȌoӡgAWhz-Ć˶iM~˾ B}XiGV(6*b>c;;qI8*z6pzl{)^g2WTLF;Gkd֭Rq,wռ.f]A7}UW-A߾IoIT Z: ¨y.zp(_-Zղj|WJ9_j44ҰRJL#>Nlgnj!=LJzs] 7?T\68KI)K:/PyUl3sTdd!Dݷl Q x>*K;L[<۸;3Ogm2U.߷l9'I$|Cgt)8.v]K]q Ys zM;o1wTmtquNy_>Tڱlawu܏S$x5Lvqlp `^~kUS2<=tZ IURD^aj$3ϕon*o=JyP}t:QJdnl%jq2iBj\LZ7z"ykF*N [qKc Z^3oaI2hfɆrpa)o C}Y/5m'^Wj!ՆtZ%;V[Tc֐㓳7Z:c1*ո!N=:R%QswHxb}zoً3J{fcXuG'Q9{~=VvEb׹t(ϪDĮc3BGzZ8ZR3hJ75n؉dhs>*'[PW~T#\Ot0S fٓ|92Pզig5sz9-m+y׻y;Y5Wmp dQ`\?^#U?a`㽣TUxsR2Ou:r$ɷmx͠*vumlF{rtt3-Est4U^?0S=~j3SS_߾l[TR\uku LY^;́w| %N iM50Tqk;;5{^à: (7ψ$?yJZߞ\nO$ wϣp2u * W`l`ؐPPBNCJU] Mw8綎H ™-!1Q63F`[42xY .7="CW=a=aQ0KO␀OPB, U#'S>g`>"(PdR Q} '$91Q̄ +83O[pjӛPrDoެ`?}T},dRA({y].1_{}c >7oQ!40 NC: bkj{-yu/5FG`{O>#̳yO>%oo4;g}"|Jha>#05vfJX+rc6EM?bhSI)}8mjJpXS 1_Dmmst=\Rt\B{|cÌ$DI|p}G1&{K K ;?;N&S~e}=0R@geVrcK%Wޮ0E{i3LS,w~ķSʀ6 {<"4lm&BEݻwzUp(Y=(nF6'vu" "$ "Sv˄͆[@V]Z34确EZfh؃OKÕz˄0fR㴵 h4C6vG %h m:JFn;e(./uT"6319{n?2Wri߁'ע=(l~uZ~{㷄jA>>:{֮ᗢE)^)P=8^o˱#;CI^kIR%A \ș_L?^?w.lo6?L H@>Dk&RMa&}Ȏfp3]s51 {"dϱ1>ePdȎ`"C, 2jTRV#'<.V ęCռZF߂uy?ɻhIqέpM^j5A4*)SN}XHd@bJd x.fϤ]KuVqtƅ_+=MStJ#,t1lTűr #60Ly,o)k pq=KOu~cqFĊ@rTyX&eY*,eKHXǭ8Be{pZ.gU)싻-&\咴TLYCF90\"ȁoţ~ ;m;d[dn*UHa T*[)e>FtP,ۓH#K @EQٟ/.D)FE$KFp$-o@s?3B +T{2.fmtg?~T]݀a>I8;~z 'ksT]a[PN]jAQ}D^jA˵*2γ,7,p虳Ԟx2/m0+|rEDEDUv#P{c?ͳź/aޢʆBSX,$ 0IrH7Ɩ{1͞\xzA#tJLQ6d15:*H_BlVٸ@qߩ%}wvU򭺃'q? ֯k{QŰ+"n '+OŘ~ioÚAp\?,T?';n6B#L,}Mɥd='q_l46zZ}HHԂgpFsk}#]8BrԆ-%)!8SמfZl06R x|LĄ%2"vI|=sZ5@U\66VCO"ȓ*%e FM=D+D .Ǭ%WBPqpϡ9 !z!nT@SwuΊKұJHu8zɠz\['9vZ@ֻ;1"]g7Axf0;tV\w砪^+lږ-նvG}Vrc/qE2L󻏫,N'$Rw&$3WA_{@/Ӂ:}V7m0}+Ni#q뾴6VUْ|F|uiezw]frNO`Kͷd2.ͻ5^Q%WC@ {:7~ؠ!k08m>I,REvA[4Bv)//w!:zmC|#A+\݁ v1l7kށ @P[jpDi ԁGk ܃.no1B5k4&hKWȧw7;ol+^Y7f?Rz TҁU=q}l$Z_z3?zݬΛ[G^ǯmoHſ,kU)%h_Z:o+[U,E&ϗo]/~%IN0//*+L}wѥ#7i{I-Ub[}E^4XlHBL2F6~NV4SiUznnfYuݏjn\ѫM6#&^$^؋$،ꑇS4c˛W%xnE ÙB+XBGэ;Q-wǔqG(Sw:,Hd:uGSwJ C|/ӇLJ,;Mߞ*MQbrztea9 poM!X JM}^y`xZoɂF?)F._*t us=7q;Q-z`)f{H۳Es0{೛Eo%&5+k@]|e9jMLLQ|`xdɳ]?HS67E=] |(qa8^&$eӂ 4p2hi*@Yjh zp}Sԃ$hn7t˥@[K-x{SV*f" hBIaQqHXt(xp!SG?Aaki4a߭' g/j6\><Q jc@pŋɋ>IQ_$Юz|׈ކo ɗ+>*DVMr,:%K_(Zj X9$ jx42bDL@_R|6^DkSErֹFz\+,yߵ%E΁V%%m{S)(KbtH\l}D)/)ӂr"K{Q؟DpZC]p\Y'i$JVBF dritשwBu#1P}MEudV/7A:V6Djۮ6fDlBFjPvukn3Iv0uw@wS%;ղ0T[Pc; 8ZAɾ7r&b I[8HK*?h k"7&ٙ (LzW PJ |:>2|O9h8w |rU/3 4JzozYrUXW]nM**s&p[ͭ؎(l}a$,/xwԤjc`WߊZ!{eFZB:p'ojo2zXB`ktIc2ޑ3bO ˠED=WܥX0BV'<- ck < T1q,Eׂ$0I6AFU#*>Bt9gQh'h=eg[k~g isƹbi {{KiF kog_cP%z[%h4+>H>Y[_=O~͵V6IBi֍/o6=yJD[7o۹=>;c_kZqDnZG<٬zS&Iu?դz5~8'q_fdζ?Ƕ'C8\׫-W>8wOc8t[4ui%mopz {q8'qoH8Y#7\hVْ?n̷KRׇ8qj-pp*1D"w̳|'ZV>R&Y4/Ly)31:;k7c!kkgvIV5%snnXQ%TH)qQ&TPVE1x zy:xV!wԗh#zDGZ,sY-r1t|TtEI ^CEDv.[.` #jŰ WLnM]y!͛߱Dln5vohq(CA4>^OT=$$ 9.Ȝ!;""5(Q1Ev!E N4 P9d#8a wҳuӞb 4iKN*,}AOy2OB(;Iێd#kF=tQ*/n> o!y 9$Z(لl"yZd]ln&:)OL)C+S9RLiCl 8k? ݴ?d 2@\aA#"3XHaж(W8 }:LlUr#n.-;r.!u\S)3/.q.<@^Bȗ8\ji C[޼8B"'I"FqU\(81IiiT c_لmp,0'4m’" X[2D>Mf[t?&JqbIhLP@p\s  G=R:' FKiiؓD4"sVY  CH=Ơp;H£Ћ}c* 4w$NhhxǗk9!s$dJB˄&dIaՁ' FHNGjZ@8}Fo4: $#nQȁx΂v&˸=nDr0_xY]o9WC Pk C-P+L1\qf^3(a+u@]"s*ƩLŘQ=S,hspw%NMȿ%NnwMGnlc.L1:y9̛lqf64Sa9bzVsp%Gb¼R >[~ 9H_^VD1˫ N;cϫ1|J9 E{ɚ o&!M%WmO)>v{jFWeqwśc^ߎ~\EAyÁu9N)sWRbnț^^*NO~|>1b?wȉgN`:/ F4ҩk:zA{:BŏgzÖf"\zCjj1MY!Mޟo܇oرMMPX5Ѯ<EVsgWj6,V}lMFӎFm:6JF_K,Fiݗs5jր!t|:?zxi>RO,ƍ~:ϭx&P*Zn~K4Nm-|8U(řX8G,/$hfEU۴J:S+hՆ2Uk];[:J7q>E Ez ^~o۝z2)^xyp!gD zoZ4Dػ!rF"qTHCζVmV.20O*n8pF6Ҵ@ EOy“H!IM bR3!xF d49Z+ ,<̯A/ FCG TJ\a(<5%y*BJSL)Z E&xS irVH Q n1|YN5TcO>oqџ3 &_͋2ξ 珋m,S{.SO 3n_#fuNV"+ϒ &JKJ*L#JCT+HI!'tC3;vr ߯r NΑܦe?i¦6lp p&6hE//ƣT$&! !0` b"3h%+\ia&VoE|pn[}_ ݌>T IUGA*~9e.5UF ;{g5կ1E4K.oNOcBǮF_7ۻق﫦cM1"0ۣӿPlZEr7[CΩR#iCp(|H{YvAЄ3[(P\~{.uSc'oN=Npltӟ:d!s'}j $ӂz:$4rOςʉE5i!pWx+ G-ag?ik'NN Oc3܂IZ޶ @JJ Z' R f0^qhslT@xh3 Q RR#R36[W~;Z-j˓`zRMF.1[W 0!?df@^PhJa$HlG] 8hZC}67S"-Cv;'7(΂RRZa33q&6Ob!뵙9qGUm3HNHCA *7TYlg0y I e֣aL7nMH 2ň7>8r؜0ptr~H!i7{c~u\@[ ox mDܼ1{ZqXk?Xdkg 3J[zө_j_Eg?zأ$ MX%}Ra(AJU2\BCh aAovp@m]-4U, )EnAB{hHx)Z=X Rp՝O&C~ܡnI܉iwCE;wh~u\ܽr8jNV0޿ȘPw\l)&?f>vk\ń Am&Mtj_2$YӜayHCܪp}Jc'>$W-6rz%- ٷ^Lߋj&+lO8$ ـO|'=wjnڨ S5&LlZ0 jkxOp'Iv7L%H~TrnZq=[ۛ%S)&E’RLSLS$(TR L”~adas.h?]jjhv:mzqam1 :x&x;cY3Tip:̵c@Wqw#5 Cj@JM^ )Ja.32UTIwN(*I,32T^i 2@hK^LT>ȼpdRKHtlMkHMaQ~>^I'Mgu镞mi'G,w-N#(-nTv=#gZhkMT\lyr{>zgV׫* ,1DXR'ے_ S eJD`37e3_/)'aKhY_oJ,1Ak40҈׌{Oi`GzL;J$m~q׷.W@\Y1"z8 8kFvaorN]_E$'7_~?gB#@n36F'?T:)x3at }LsY&c 0n,V6hwV1F&V)m m#06}jrMdנŮ[}d'jLeUP-(V xeҤ D JMísn"aٶӶ2Ϲ@N(hQd2cy)9bXY.K]IʙrlF{~rJP=2u6DYP~s{CA)a%''*!ܞ9QFiu?;sȠč۳9n\ u ~T3oћ&'0Kn]usA:ZT4;<@L%pJ !Pז[m@.mB ƔqĻ41'C͛*$i+3P7t6lǬ `.Rɽ +{{ꔥ4 0#tw#Mv^^]D Ջ]pQujT`q`g[Nb3!]i{ܕu֕6Jۄp-S Wwj Tcl,k$`h|1lbb)E7\Sg192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 15:54:55 crc kubenswrapper[4880]: I0126 15:54:55.835359 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43768->192.168.126.11:17697: read: connection reset by peer" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.072559 4880 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 26 15:54:56 crc kubenswrapper[4880]: E0126 15:54:56.270291 4880 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.298263 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 02:50:43.352001719 +0000 UTC Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.803272 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.805617 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8" exitCode=255 Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.805796 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8"} Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.806217 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.807625 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.807667 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.807681 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:54:56 crc kubenswrapper[4880]: I0126 15:54:56.808535 4880 scope.go:117] "RemoveContainer" containerID="b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8" Jan 26 15:54:57 crc kubenswrapper[4880]: E0126 15:54:57.255581 4880 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.188e52ebb922a1d4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 15:54:35.98003042 +0000 UTC m=+1.465759137,LastTimestamp:2026-01-26 15:54:35.98003042 +0000 UTC m=+1.465759137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.298810 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 12:19:52.123436578 +0000 UTC Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.812136 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.815332 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7"} Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.821994 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.823547 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.823800 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:54:57 crc kubenswrapper[4880]: I0126 15:54:57.824352 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:54:58 crc kubenswrapper[4880]: I0126 15:54:58.180897 4880 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" start-of-body= Jan 26 15:54:58 crc kubenswrapper[4880]: I0126 15:54:58.181002 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded" Jan 26 15:54:58 crc kubenswrapper[4880]: I0126 15:54:58.315720 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 05:32:41.775222102 +0000 UTC Jan 26 15:54:58 crc kubenswrapper[4880]: E0126 15:54:58.709177 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="7s" Jan 26 15:54:59 crc kubenswrapper[4880]: I0126 15:54:59.005791 4880 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 26 15:54:59 crc kubenswrapper[4880]: I0126 15:54:59.005855 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 26 15:54:59 crc kubenswrapper[4880]: I0126 15:54:59.316669 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 04:02:04.761893691 +0000 UTC Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.317641 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 21:38:36.895720539 +0000 UTC Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.821477 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.821657 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.823023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.823073 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.823085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:00 crc kubenswrapper[4880]: I0126 15:55:00.834634 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.318313 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 16:36:59.606035228 +0000 UTC Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.371910 4880 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.372012 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.827774 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.828917 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.828966 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:01 crc kubenswrapper[4880]: I0126 15:55:01.828978 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:02 crc kubenswrapper[4880]: I0126 15:55:02.319286 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 00:13:59.482247457 +0000 UTC Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.187551 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.187782 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.187953 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.189081 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.189116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.189126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.193164 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.320514 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 00:55:13.370294562 +0000 UTC Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.833002 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.834351 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.834396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:03 crc kubenswrapper[4880]: I0126 15:55:03.834409 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.001560 4880 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.002263 4880 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 15:55:04 crc kubenswrapper[4880]: E0126 15:55:04.003997 4880 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.006930 4880 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.008530 4880 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.208155 4880 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 15:55:04 crc kubenswrapper[4880]: I0126 15:55:04.321358 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 06:59:18.086315172 +0000 UTC Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.076058 4880 apiserver.go:52] "Watching apiserver" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.080753 4880 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.081405 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.082110 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.082196 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.082216 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.082225 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.082504 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.083034 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.083069 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.083378 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.083471 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.085002 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.085394 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.086489 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.086565 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.086741 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.086933 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.087064 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.087260 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.089855 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.099867 4880 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.120875 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.120963 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121002 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121029 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121054 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121095 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121118 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121138 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121160 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121184 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121206 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121232 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121257 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121272 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121298 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121318 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121340 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121670 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121740 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.121924 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122178 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122232 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122218 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122234 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122248 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122341 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122799 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122576 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122617 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122629 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122643 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122968 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.122998 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123030 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123050 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123127 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123146 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123171 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123199 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123218 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123242 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123265 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123304 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123324 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123348 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123367 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123391 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123413 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123460 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123488 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123511 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123535 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123558 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123597 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123620 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123647 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123669 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123689 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123710 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123732 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123756 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123775 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123808 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123833 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123855 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123879 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123898 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123917 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123935 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123954 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123975 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123999 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124021 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124041 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124063 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124094 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124120 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124143 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124163 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124188 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124227 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124289 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124310 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124328 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124346 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124367 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124391 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124413 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124460 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124485 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124506 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124524 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124543 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123047 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124565 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123298 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124578 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124592 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124617 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124651 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124677 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124715 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124738 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124759 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124824 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124842 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124970 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125005 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125064 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125088 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125107 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125126 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125143 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125183 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125204 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125224 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125244 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125262 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125284 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125318 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125338 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125359 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125377 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125400 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125423 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125466 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125488 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125507 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125527 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125571 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125597 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125636 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125664 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125687 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125707 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125725 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125746 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125775 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125798 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125841 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125869 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125889 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125914 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125936 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125958 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125983 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126005 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126050 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126072 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126097 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126119 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126148 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126169 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126198 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126239 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126299 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126322 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126342 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126389 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126416 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126454 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126481 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126502 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126540 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126569 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126599 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126632 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126698 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126720 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126740 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126769 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126791 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126814 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126839 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126946 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126974 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127020 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127198 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127236 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127284 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127307 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127360 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127384 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127414 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127491 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127567 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127591 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127611 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127633 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127651 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127674 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127696 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127716 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127745 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127770 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127794 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127821 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127845 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127870 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127895 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127921 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127941 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127964 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127985 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128033 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128055 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128117 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128166 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128189 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128265 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128287 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128377 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128409 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128456 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128496 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128520 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128542 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128566 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128586 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128608 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128630 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128649 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128673 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128709 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128734 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128823 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128838 4880 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128852 4880 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128863 4880 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128875 4880 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128887 4880 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128899 4880 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128911 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128922 4880 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128933 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128943 4880 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128955 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128967 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128978 4880 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128989 4880 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123466 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.123768 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124011 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124278 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124381 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.124858 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125428 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125658 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125925 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.125990 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126470 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126620 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126525 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.126922 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127125 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127452 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127541 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.127820 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128071 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.128293 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.130006 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.131757 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132037 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132069 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132478 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132621 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132702 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.132922 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.133352 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.133512 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.134519 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.135094 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.135230 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.135357 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.135874 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.136207 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.136513 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.136518 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.136969 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137124 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137308 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137628 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137662 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137696 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137702 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.137756 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138023 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138085 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138142 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138413 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138510 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138558 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138816 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.138870 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.139203 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.139373 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.139549 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.140013 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.140476 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.140742 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.141213 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.141724 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.141786 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.142171 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.142418 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.142650 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.142677 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.144203 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.144295 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.144823 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.144857 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.145342 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.145817 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.147119 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.147576 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.152454 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.152460 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.152851 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.153025 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:05.652976375 +0000 UTC m=+31.138705082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.153556 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154018 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154104 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154365 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154419 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154790 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.154855 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.155415 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.155586 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.156091 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.157012 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.157605 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.158188 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.158956 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.162421 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.162749 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.163793 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164084 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164196 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164213 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164381 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164827 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164711 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.165104 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.165170 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.166129 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.166653 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.167068 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.167075 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.167169 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.167453 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.168903 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.169166 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.164719 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.188040 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.188584 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.188930 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.189424 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.189663 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.189848 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.190140 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.190343 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.190603 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.158964 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.190706 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.190796 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.191190 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.191526 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.194337 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.194785 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.195903 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.196064 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.196629 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.196719 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.196747 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.191265 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.197384 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.197791 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.198196 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.198369 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.198518 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.198572 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.199553 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.199528 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.199757 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.199873 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.199891 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:05.699866056 +0000 UTC m=+31.185594763 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.200161 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.201004 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.201365 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.201668 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.201970 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.202167 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.202739 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.203011 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.203850 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.204021 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.205097 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.205624 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.206423 4880 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.206466 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.206495 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.206519 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.206603 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:05.706581275 +0000 UTC m=+31.192309982 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.206899 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.207790 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210124 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210399 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210565 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210763 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210866 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210937 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.210932 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.211074 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.212633 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.153593 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.216209 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.219552 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.219804 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.220483 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.220587 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.221149 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.221267 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:05.721240852 +0000 UTC m=+31.206969559 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.222100 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.229847 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.229903 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.230859 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.232536 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.232578 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.232599 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.232679 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:05.732652223 +0000 UTC m=+31.218381000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233651 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233687 4880 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233700 4880 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233710 4880 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233721 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233732 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233756 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233770 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233781 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233792 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233806 4880 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233823 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233845 4880 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233856 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233865 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233874 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233883 4880 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233900 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233921 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233937 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.233959 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234014 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234031 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234043 4880 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234054 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234097 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234137 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234158 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234186 4880 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234220 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234233 4880 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234244 4880 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234265 4880 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234284 4880 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234526 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234693 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234703 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234712 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234722 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234731 4880 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234763 4880 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234774 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234786 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234796 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234808 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234838 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234849 4880 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234858 4880 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234867 4880 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234876 4880 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234885 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234894 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234955 4880 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.234969 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235011 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235022 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235031 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235042 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235049 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235058 4880 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235066 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235098 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235108 4880 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235127 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235137 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235146 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235179 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235189 4880 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235201 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235209 4880 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235220 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235264 4880 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235276 4880 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235302 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235364 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235375 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235384 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235411 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235422 4880 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235464 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235675 4880 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235691 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235838 4880 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235865 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235909 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235922 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235934 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235945 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235954 4880 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235987 4880 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.235997 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236006 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236014 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236023 4880 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236031 4880 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236062 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236074 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236083 4880 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236092 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236100 4880 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236108 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236116 4880 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236147 4880 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236242 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236255 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236264 4880 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236272 4880 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236309 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236324 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236336 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236344 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236354 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236406 4880 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236417 4880 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236425 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236462 4880 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236474 4880 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236483 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236492 4880 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236500 4880 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236510 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236543 4880 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236552 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236563 4880 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236573 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236581 4880 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236590 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236622 4880 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236632 4880 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236646 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236663 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236709 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236728 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236744 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236759 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236791 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236802 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236811 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236832 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236842 4880 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236905 4880 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236914 4880 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236922 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236968 4880 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.236989 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237004 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237047 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237071 4880 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237089 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237100 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237137 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237153 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237162 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237171 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237181 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237213 4880 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237223 4880 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237231 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237241 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237250 4880 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237259 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237295 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237311 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237320 4880 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237328 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237336 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237344 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237375 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237385 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237393 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237471 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237596 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.229957 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.237882 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.244834 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.250808 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.252269 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.256997 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.258525 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.258606 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.258642 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.258666 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.261270 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.266046 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.270730 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.279181 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.291951 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.297620 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.321746 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 09:18:05.390512005 +0000 UTC Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.329249 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338064 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338126 4880 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338153 4880 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338172 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338195 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338216 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338237 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338263 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.338291 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.352166 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.371286 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:54:55Z\\\",\\\"message\\\":\\\"W0126 15:54:43.999550 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0126 15:54:44.000048 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769442884 cert, and key in /tmp/serving-cert-62186932/serving-signer.crt, /tmp/serving-cert-62186932/serving-signer.key\\\\nI0126 15:54:44.919856 1 observer_polling.go:159] Starting file observer\\\\nW0126 15:54:45.203139 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0126 15:54:45.203362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 15:54:45.237830 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-62186932/tls.crt::/tmp/serving-cert-62186932/tls.key\\\\\\\"\\\\nF0126 15:54:55.769431 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.394414 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.412479 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.432485 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.456643 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 15:55:05 crc kubenswrapper[4880]: W0126 15:55:05.507329 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a1356eac0ee8ad3a245183464bb15c02023654c036157b13de90551466314dcd WatchSource:0}: Error finding container a1356eac0ee8ad3a245183464bb15c02023654c036157b13de90551466314dcd: Status 404 returned error can't find the container with id a1356eac0ee8ad3a245183464bb15c02023654c036157b13de90551466314dcd Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.742945 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743202 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:06.743165445 +0000 UTC m=+32.228894152 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.743491 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.743538 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.743571 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.743608 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743694 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743708 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743767 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743780 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:06.74375737 +0000 UTC m=+32.229486127 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743794 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743811 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743826 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743855 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743909 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743910 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:06.743794301 +0000 UTC m=+32.229523038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.743981 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:06.743953624 +0000 UTC m=+32.229682371 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.744016 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:06.744010386 +0000 UTC m=+32.229739093 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.928758 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.929420 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.933070 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" exitCode=255 Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.933165 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.933249 4880 scope.go:117] "RemoveContainer" containerID="b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.934225 4880 scope.go:117] "RemoveContainer" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" Jan 26 15:55:05 crc kubenswrapper[4880]: E0126 15:55:05.934517 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.941564 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a1356eac0ee8ad3a245183464bb15c02023654c036157b13de90551466314dcd"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.948938 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.949064 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.949105 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"96321ff5ca2e6732b9b8a50e91f9c6e0d4db52bc27fd502b00b0bbd35c3e3574"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.952654 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.952719 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"317abf530fc39058e94891ae086456dcabfb4a80ea589bb88198fea83cf068aa"} Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.960770 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:54:55Z\\\",\\\"message\\\":\\\"W0126 15:54:43.999550 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0126 15:54:44.000048 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769442884 cert, and key in /tmp/serving-cert-62186932/serving-signer.crt, /tmp/serving-cert-62186932/serving-signer.key\\\\nI0126 15:54:44.919856 1 observer_polling.go:159] Starting file observer\\\\nW0126 15:54:45.203139 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0126 15:54:45.203362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 15:54:45.237830 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-62186932/tls.crt::/tmp/serving-cert-62186932/tls.key\\\\\\\"\\\\nF0126 15:54:55.769431 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:05 crc kubenswrapper[4880]: I0126 15:55:05.982810 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.005812 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.022338 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.045215 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.062211 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.163885 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.185378 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.208524 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.213758 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.214008 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.222278 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.222953 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.224615 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.225528 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.226663 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.227241 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.227965 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.229395 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.230299 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.231619 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.232266 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.232637 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.233875 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.234528 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.235390 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.236802 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.237484 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.238801 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.239268 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.240051 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.241521 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.242203 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.243595 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.244148 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.245583 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.246179 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.247230 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.248883 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.251536 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.252519 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.253625 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.256740 4880 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.257118 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.259667 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.260671 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.262340 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.264538 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.265518 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.266258 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.267213 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.268185 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.268953 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.273156 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.275637 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.326919 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 19:18:25.167457537 +0000 UTC Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.437783 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.439711 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.443059 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.445477 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.448418 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.449881 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.451844 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.453101 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.454524 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.456650 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.457632 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.521862 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:54:55Z\\\",\\\"message\\\":\\\"W0126 15:54:43.999550 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0126 15:54:44.000048 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769442884 cert, and key in /tmp/serving-cert-62186932/serving-signer.crt, /tmp/serving-cert-62186932/serving-signer.key\\\\nI0126 15:54:44.919856 1 observer_polling.go:159] Starting file observer\\\\nW0126 15:54:45.203139 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0126 15:54:45.203362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 15:54:45.237830 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-62186932/tls.crt::/tmp/serving-cert-62186932/tls.key\\\\\\\"\\\\nF0126 15:54:55.769431 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.556820 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.775847 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.775961 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.775995 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.776017 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.776037 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776136 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:08.776101564 +0000 UTC m=+34.261830271 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776177 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776218 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776253 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776262 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776269 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776278 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776241 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776336 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776325 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:08.776308839 +0000 UTC m=+34.262037546 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776405 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:08.776394511 +0000 UTC m=+34.262123218 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776418 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:08.776410991 +0000 UTC m=+34.262139698 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.776484 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:08.776456742 +0000 UTC m=+34.262185449 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.792824 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.803556 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.815701 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.829850 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.844349 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.859381 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.926298 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5a8b4ee14252b3f2f9242ac5ba1971f28820eb0ad962850f49c83f17d03fac8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:54:55Z\\\",\\\"message\\\":\\\"W0126 15:54:43.999550 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0126 15:54:44.000048 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769442884 cert, and key in /tmp/serving-cert-62186932/serving-signer.crt, /tmp/serving-cert-62186932/serving-signer.key\\\\nI0126 15:54:44.919856 1 observer_polling.go:159] Starting file observer\\\\nW0126 15:54:45.203139 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0126 15:54:45.203362 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 15:54:45.237830 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-62186932/tls.crt::/tmp/serving-cert-62186932/tls.key\\\\\\\"\\\\nF0126 15:54:55.769431 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.942291 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.957125 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.957127 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.959551 4880 scope.go:117] "RemoveContainer" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" Jan 26 15:55:06 crc kubenswrapper[4880]: E0126 15:55:06.959752 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.974408 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:06 crc kubenswrapper[4880]: I0126 15:55:06.991238 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.005397 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:07Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.019561 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:07Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.033995 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:07Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.047592 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:07Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.062348 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:07Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.214591 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.214746 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:07 crc kubenswrapper[4880]: E0126 15:55:07.214873 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:07 crc kubenswrapper[4880]: E0126 15:55:07.214982 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.407828 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 12:26:37.057751782 +0000 UTC Jan 26 15:55:07 crc kubenswrapper[4880]: I0126 15:55:07.453249 4880 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.214466 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.214640 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.379219 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.384798 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.392626 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.511218 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 07:56:45.124443529 +0000 UTC Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.517771 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.530941 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.550318 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.563945 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.579630 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.597611 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.613617 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.634816 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.706264 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.874758 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.874839 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.874860 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.874931 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:08 crc kubenswrapper[4880]: I0126 15:55:08.874990 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.875229 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.875253 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.875269 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.897738 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.898419 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:12.898374936 +0000 UTC m=+38.384103643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.899486 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:12.899464201 +0000 UTC m=+38.385192918 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.899822 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.900353 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:12.900303192 +0000 UTC m=+38.386031899 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.900747 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.900926 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.901093 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:08 crc kubenswrapper[4880]: E0126 15:55:08.901300 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:12.901270255 +0000 UTC m=+38.386998972 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.025366 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:08Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: E0126 15:55:09.034088 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:13.03404596 +0000 UTC m=+38.519774667 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.153822 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: E0126 15:55:09.169292 4880 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.214027 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.214152 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:09 crc kubenswrapper[4880]: E0126 15:55:09.214800 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:09 crc kubenswrapper[4880]: E0126 15:55:09.214798 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.376178 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.448285 4880 csr.go:261] certificate signing request csr-kplfm is approved, waiting to be issued Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.513002 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:27:22.275638551 +0000 UTC Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.546923 4880 csr.go:257] certificate signing request csr-kplfm is issued Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.548176 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.569565 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.578652 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2ks4h"] Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.579117 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.582516 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.582859 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.585299 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.585456 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.595844 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.610664 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.625112 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.645286 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.659322 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.677745 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.693881 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.707790 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512ee353-65e1-404a-9f3c-ee314b873f5e-host\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.707890 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/512ee353-65e1-404a-9f3c-ee314b873f5e-serviceca\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.707922 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlwqg\" (UniqueName: \"kubernetes.io/projected/512ee353-65e1-404a-9f3c-ee314b873f5e-kube-api-access-rlwqg\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.710146 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.723987 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.737107 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.808733 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/512ee353-65e1-404a-9f3c-ee314b873f5e-serviceca\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.808793 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlwqg\" (UniqueName: \"kubernetes.io/projected/512ee353-65e1-404a-9f3c-ee314b873f5e-kube-api-access-rlwqg\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.808822 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512ee353-65e1-404a-9f3c-ee314b873f5e-host\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.808884 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/512ee353-65e1-404a-9f3c-ee314b873f5e-host\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.812520 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/512ee353-65e1-404a-9f3c-ee314b873f5e-serviceca\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.845413 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlwqg\" (UniqueName: \"kubernetes.io/projected/512ee353-65e1-404a-9f3c-ee314b873f5e-kube-api-access-rlwqg\") pod \"node-ca-2ks4h\" (UID: \"512ee353-65e1-404a-9f3c-ee314b873f5e\") " pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: I0126 15:55:09.895304 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2ks4h" Jan 26 15:55:09 crc kubenswrapper[4880]: W0126 15:55:09.916008 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod512ee353_65e1_404a_9f3c_ee314b873f5e.slice/crio-da01f28a7f22ef46aadad8060bacc8b42e211f33531f1565f49ef1046706b7e5 WatchSource:0}: Error finding container da01f28a7f22ef46aadad8060bacc8b42e211f33531f1565f49ef1046706b7e5: Status 404 returned error can't find the container with id da01f28a7f22ef46aadad8060bacc8b42e211f33531f1565f49ef1046706b7e5 Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.034678 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2ks4h" event={"ID":"512ee353-65e1-404a-9f3c-ee314b873f5e","Type":"ContainerStarted","Data":"da01f28a7f22ef46aadad8060bacc8b42e211f33531f1565f49ef1046706b7e5"} Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.079139 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-glj8k"] Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.079662 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.081301 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mpnc5"] Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.081747 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-h6f8z"] Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.081973 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.082577 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.082642 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.082853 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.082588 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.083361 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.083628 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.085904 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.086092 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.086246 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.086592 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.086783 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.086952 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.087074 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.087265 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.197923 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213275 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213349 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/db202b43-2be9-4b6c-87b7-b2c5ef2df984-rootfs\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213378 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d4mt\" (UniqueName: \"kubernetes.io/projected/db202b43-2be9-4b6c-87b7-b2c5ef2df984-kube-api-access-6d4mt\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213399 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgqzb\" (UniqueName: \"kubernetes.io/projected/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-kube-api-access-xgqzb\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213454 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-multus\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213480 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-daemon-config\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213538 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-hostroot\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213565 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db202b43-2be9-4b6c-87b7-b2c5ef2df984-mcd-auth-proxy-config\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213600 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-cnibin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213613 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-kubelet\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213636 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-etc-kubernetes\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213702 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-hosts-file\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213733 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-bin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213779 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-os-release\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213800 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-cni-binary-copy\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213889 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-netns\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213940 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-conf-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213962 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/db202b43-2be9-4b6c-87b7-b2c5ef2df984-proxy-tls\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.213979 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-socket-dir-parent\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.214049 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vz2h\" (UniqueName: \"kubernetes.io/projected/ddcd016a-fdad-4580-adaf-30451ac85f0e-kube-api-access-5vz2h\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.214068 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-system-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.214088 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-k8s-cni-cncf-io\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.214121 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-multus-certs\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.216711 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:10 crc kubenswrapper[4880]: E0126 15:55:10.217003 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.217640 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.247154 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.314739 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-hosts-file\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.314810 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-bin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.314920 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-os-release\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.314957 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-cni-binary-copy\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315000 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-netns\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315081 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-conf-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315112 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/db202b43-2be9-4b6c-87b7-b2c5ef2df984-proxy-tls\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315138 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-socket-dir-parent\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315170 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vz2h\" (UniqueName: \"kubernetes.io/projected/ddcd016a-fdad-4580-adaf-30451ac85f0e-kube-api-access-5vz2h\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315234 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-system-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315261 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-k8s-cni-cncf-io\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315319 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-multus-certs\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315348 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315375 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/db202b43-2be9-4b6c-87b7-b2c5ef2df984-rootfs\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315404 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d4mt\" (UniqueName: \"kubernetes.io/projected/db202b43-2be9-4b6c-87b7-b2c5ef2df984-kube-api-access-6d4mt\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315428 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgqzb\" (UniqueName: \"kubernetes.io/projected/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-kube-api-access-xgqzb\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315472 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-daemon-config\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315499 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-multus\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315544 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-hostroot\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315592 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db202b43-2be9-4b6c-87b7-b2c5ef2df984-mcd-auth-proxy-config\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315621 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-cnibin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315662 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-kubelet\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315689 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-etc-kubernetes\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315775 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-etc-kubernetes\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315858 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-hosts-file\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.315901 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-bin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.316345 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-os-release\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.316566 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.317367 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-cni-binary-copy\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.317425 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-netns\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.317525 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-conf-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.318630 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/db202b43-2be9-4b6c-87b7-b2c5ef2df984-rootfs\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.318570 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-cni-multus\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.318865 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-cnibin\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.318898 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-var-lib-kubelet\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319023 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-system-cni-dir\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319125 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-socket-dir-parent\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319391 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-k8s-cni-cncf-io\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319426 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-host-run-multus-certs\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319469 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ddcd016a-fdad-4580-adaf-30451ac85f0e-hostroot\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.319467 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/db202b43-2be9-4b6c-87b7-b2c5ef2df984-mcd-auth-proxy-config\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.320984 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ddcd016a-fdad-4580-adaf-30451ac85f0e-multus-daemon-config\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.322416 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/db202b43-2be9-4b6c-87b7-b2c5ef2df984-proxy-tls\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.420880 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d4mt\" (UniqueName: \"kubernetes.io/projected/db202b43-2be9-4b6c-87b7-b2c5ef2df984-kube-api-access-6d4mt\") pod \"machine-config-daemon-glj8k\" (UID: \"db202b43-2be9-4b6c-87b7-b2c5ef2df984\") " pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.499015 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vz2h\" (UniqueName: \"kubernetes.io/projected/ddcd016a-fdad-4580-adaf-30451ac85f0e-kube-api-access-5vz2h\") pod \"multus-h6f8z\" (UID: \"ddcd016a-fdad-4580-adaf-30451ac85f0e\") " pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.519573 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgqzb\" (UniqueName: \"kubernetes.io/projected/fa2af6c3-36e8-40a6-816b-52d5d64fe9ad-kube-api-access-xgqzb\") pod \"node-resolver-mpnc5\" (UID: \"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\") " pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.520329 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:26:49.945868947 +0000 UTC Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.547855 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.548583 4880 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-26 15:50:09 +0000 UTC, rotation deadline is 2026-12-01 06:23:10.260924727 +0000 UTC Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.548611 4880 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7406h27m59.712316631s for next certificate rotation Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.568698 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.623972 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.628023 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-b2pvv"] Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.628871 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: W0126 15:55:10.630558 4880 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 26 15:55:10 crc kubenswrapper[4880]: E0126 15:55:10.630628 4880 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.631336 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.644697 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.672053 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.687081 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.698986 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:55:10 crc kubenswrapper[4880]: W0126 15:55:10.713494 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb202b43_2be9_4b6c_87b7_b2c5ef2df984.slice/crio-c69de85ea16e0e616f556d6288b1538a77a148cd1f70659af0fbdce40ae92dc0 WatchSource:0}: Error finding container c69de85ea16e0e616f556d6288b1538a77a148cd1f70659af0fbdce40ae92dc0: Status 404 returned error can't find the container with id c69de85ea16e0e616f556d6288b1538a77a148cd1f70659af0fbdce40ae92dc0 Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.714751 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.721519 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-h6f8z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.721900 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-binary-copy\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.721954 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-system-cni-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.722007 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.722036 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq76j\" (UniqueName: \"kubernetes.io/projected/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-kube-api-access-kq76j\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.722062 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cnibin\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.722118 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-os-release\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.722145 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.737796 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.742307 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mpnc5" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.754896 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.777521 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: W0126 15:55:10.782797 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddcd016a_fdad_4580_adaf_30451ac85f0e.slice/crio-4c0089b672da72c72b740ccc5043406ff86a8a98eaf331dff7cd26dec6089052 WatchSource:0}: Error finding container 4c0089b672da72c72b740ccc5043406ff86a8a98eaf331dff7cd26dec6089052: Status 404 returned error can't find the container with id 4c0089b672da72c72b740ccc5043406ff86a8a98eaf331dff7cd26dec6089052 Jan 26 15:55:10 crc kubenswrapper[4880]: W0126 15:55:10.786751 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa2af6c3_36e8_40a6_816b_52d5d64fe9ad.slice/crio-b44e1c1d1745caf2b9f4845f4e37f8153a1c8c86e0014ef0ff11fcd68aa380cc WatchSource:0}: Error finding container b44e1c1d1745caf2b9f4845f4e37f8153a1c8c86e0014ef0ff11fcd68aa380cc: Status 404 returned error can't find the container with id b44e1c1d1745caf2b9f4845f4e37f8153a1c8c86e0014ef0ff11fcd68aa380cc Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.797077 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823557 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823610 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq76j\" (UniqueName: \"kubernetes.io/projected/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-kube-api-access-kq76j\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823645 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cnibin\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823676 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-os-release\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823713 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823759 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-binary-copy\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823788 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-system-cni-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.823902 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-system-cni-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.824531 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-os-release\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.824637 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.824703 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-tuning-conf-dir\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.824786 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.824608 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cnibin\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.825386 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-cni-binary-copy\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.848550 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.861148 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq76j\" (UniqueName: \"kubernetes.io/projected/8dbe43e2-b41e-4bbe-8bb3-c3389816a067-kube-api-access-kq76j\") pod \"multus-additional-cni-plugins-b2pvv\" (UID: \"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\") " pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.865858 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.882212 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.901853 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.923609 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.942902 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.966249 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:10 crc kubenswrapper[4880]: I0126 15:55:10.984470 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:10Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.004253 4880 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.031585 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.031669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.031865 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.032050 4880 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.043833 4880 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.043947 4880 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.045725 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerStarted","Data":"b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.045786 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerStarted","Data":"4c0089b672da72c72b740ccc5043406ff86a8a98eaf331dff7cd26dec6089052"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.050326 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.050393 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"c69de85ea16e0e616f556d6288b1538a77a148cd1f70659af0fbdce40ae92dc0"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.051691 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2ks4h" event={"ID":"512ee353-65e1-404a-9f3c-ee314b873f5e","Type":"ContainerStarted","Data":"ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.056852 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mpnc5" event={"ID":"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad","Type":"ContainerStarted","Data":"b44e1c1d1745caf2b9f4845f4e37f8153a1c8c86e0014ef0ff11fcd68aa380cc"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.073598 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.079291 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.079346 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.079357 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.079378 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.079394 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.085581 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2p7jp"] Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.095514 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.096536 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.098175 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.103050 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.110938 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.111333 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.111412 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.111413 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.111757 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.123151 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.141109 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.141185 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.141211 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.141269 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.141292 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.142393 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.213679 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.213736 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.213930 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.214067 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227750 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227799 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85xjf\" (UniqueName: \"kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227822 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227840 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227858 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227874 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227896 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227931 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227947 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.227979 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228004 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228025 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228040 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228063 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228078 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228275 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228356 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228387 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228420 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.228532 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.306310 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.312782 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.320211 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.320270 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.320280 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.320297 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.320307 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.329891 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.329950 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330004 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330029 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330075 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330101 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330170 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330166 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330175 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330273 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330202 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330540 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330319 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330345 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330287 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330641 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330861 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.330989 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331106 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331165 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331196 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331232 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331234 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331253 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331167 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331278 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331419 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331487 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331582 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331652 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331739 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331769 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85xjf\" (UniqueName: \"kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.332549 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.331556 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.333368 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.333482 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.333773 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.334848 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.336957 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.336998 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.372202 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85xjf\" (UniqueName: \"kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf\") pod \"ovnkube-node-2p7jp\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.531384 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.531475 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.531548 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 11:41:10.056563172 +0000 UTC Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.542727 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.542803 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.542818 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.542839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.542851 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.625740 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.626848 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.627203 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.627586 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.637650 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.655505 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.655549 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.655561 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.655580 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.655595 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.673561 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: W0126 15:55:11.675949 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dbe43e2_b41e_4bbe_8bb3_c3389816a067.slice/crio-0287a96d6be7d17e454bb518fe736def3764ec5b40c8fc7b65c9b76e3a582061 WatchSource:0}: Error finding container 0287a96d6be7d17e454bb518fe736def3764ec5b40c8fc7b65c9b76e3a582061: Status 404 returned error can't find the container with id 0287a96d6be7d17e454bb518fe736def3764ec5b40c8fc7b65c9b76e3a582061 Jan 26 15:55:11 crc kubenswrapper[4880]: W0126 15:55:11.677346 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc377ad66_a182_4cbf_9b29_253f3e547118.slice/crio-7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf WatchSource:0}: Error finding container 7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf: Status 404 returned error can't find the container with id 7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.678813 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: E0126 15:55:11.679029 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.680671 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.680709 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.680723 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.680741 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.680755 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.696272 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.717481 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.735036 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.749865 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.767054 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.783882 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.783953 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.783970 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.783992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.784008 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.785561 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.801533 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.817103 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.870082 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.888361 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.888427 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.888478 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.888499 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.888515 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:11Z","lastTransitionTime":"2026-01-26T15:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:11 crc kubenswrapper[4880]: I0126 15:55:11.914240 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:11Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.024645 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.024697 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.024714 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.024733 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.024750 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.038250 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.117813 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerStarted","Data":"0287a96d6be7d17e454bb518fe736def3764ec5b40c8fc7b65c9b76e3a582061"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.118993 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mpnc5" event={"ID":"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad","Type":"ContainerStarted","Data":"8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140208 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140346 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140362 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140371 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.140395 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.147125 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.212541 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.217848 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.217969 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.248887 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.248992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.249028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.249083 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.249118 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.309853 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.366897 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.366998 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.367033 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.367062 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.367077 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.378483 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.398868 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.471315 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.471384 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.471396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.471416 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.471428 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.481593 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.532125 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 23:15:49.253734714 +0000 UTC Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.558171 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.579560 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.579632 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.579674 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.579714 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.579740 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.755958 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.755992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.756000 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.756033 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.756046 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.760962 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.934940 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.935086 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.935134 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.935160 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.935322 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.935346 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.935359 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.935419 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:20.935401595 +0000 UTC m=+46.421130312 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.935938 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:12Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.936172 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:20.936159493 +0000 UTC m=+46.421888200 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.936237 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.936270 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:20.936261405 +0000 UTC m=+46.421990112 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.936306 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.936333 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:20.936325537 +0000 UTC m=+46.422054244 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.937978 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.938012 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.938023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.938039 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.938051 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:12Z","lastTransitionTime":"2026-01-26T15:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.957571 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:12 crc kubenswrapper[4880]: I0126 15:55:12.958274 4880 scope.go:117] "RemoveContainer" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" Jan 26 15:55:12 crc kubenswrapper[4880]: E0126 15:55:12.958481 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.454805 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.455027 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.455316 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.455391 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.455322 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.455486 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.455590 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.560501 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.560730 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:21.560674687 +0000 UTC m=+47.046403394 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.560235 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 06:38:10.920693389 +0000 UTC Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.583024 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.583089 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.583108 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.583127 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.583140 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:13Z","lastTransitionTime":"2026-01-26T15:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.608275 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerStarted","Data":"8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4"} Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.611112 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba" exitCode=0 Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.612464 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba"} Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.612763 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:13 crc kubenswrapper[4880]: E0126 15:55:13.612955 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.627090 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.670484 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.685632 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.685687 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.685697 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.685713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.685741 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:13Z","lastTransitionTime":"2026-01-26T15:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.734795 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.778081 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.800786 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.807053 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.807138 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.807153 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.807171 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.807211 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:13Z","lastTransitionTime":"2026-01-26T15:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.822467 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.903368 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.910905 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.910976 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.910990 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.911011 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:13 crc kubenswrapper[4880]: I0126 15:55:13.911029 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:13Z","lastTransitionTime":"2026-01-26T15:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.028960 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:13Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.030943 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.031008 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.031022 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.031042 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.031055 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.049101 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.075756 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.330404 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.337155 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.337194 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.337207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.337224 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.337248 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.394282 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.430153 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.443981 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.444069 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.444085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.444143 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.444164 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.454662 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.485324 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.503017 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.519610 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.535577 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.548879 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.548972 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.548988 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.549006 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.549018 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.553845 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.561654 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 02:07:35.256720836 +0000 UTC Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.572355 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.589136 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.612664 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.624366 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.651896 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.651931 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.651942 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.651957 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.651968 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.811103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.811167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.811182 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.811221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.811236 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:14Z","lastTransitionTime":"2026-01-26T15:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:14 crc kubenswrapper[4880]: I0126 15:55:14.833241 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:14Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.284101 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.284952 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:15 crc kubenswrapper[4880]: E0126 15:55:15.285216 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.285417 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:15 crc kubenswrapper[4880]: E0126 15:55:15.285615 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.287289 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.287427 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.287801 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.287950 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.288506 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:15Z","lastTransitionTime":"2026-01-26T15:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:15 crc kubenswrapper[4880]: E0126 15:55:15.284558 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.381965 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:15Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.392323 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.392387 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.392401 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.392420 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.392497 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:15Z","lastTransitionTime":"2026-01-26T15:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.499480 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.499523 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.499546 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.499570 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.499590 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:15Z","lastTransitionTime":"2026-01-26T15:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.562991 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 18:45:03.660085186 +0000 UTC Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.796976 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.797035 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.797047 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.797068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.797093 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:15Z","lastTransitionTime":"2026-01-26T15:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.882366 4880 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.885789 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ovn-kubernetes/pods/ovnkube-node-2p7jp/status\": read tcp 38.102.83.188:60166->38.102.83.188:6443: use of closed network connection" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.906574 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.906650 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.906679 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.906709 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:15 crc kubenswrapper[4880]: I0126 15:55:15.906733 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:15Z","lastTransitionTime":"2026-01-26T15:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.211299 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.211374 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.211390 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.211421 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.211493 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:16Z","lastTransitionTime":"2026-01-26T15:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.383627 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.383691 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.383726 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.383776 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.383798 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:16Z","lastTransitionTime":"2026-01-26T15:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.493460 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.493596 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.493621 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.493657 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.493686 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:16Z","lastTransitionTime":"2026-01-26T15:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.563207 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 12:13:03.82721778 +0000 UTC Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.712552 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.712589 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.712600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.712614 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.712624 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:16Z","lastTransitionTime":"2026-01-26T15:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.722697 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.759191 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.782412 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.845633 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.845669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.845678 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.845692 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.845701 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:16Z","lastTransitionTime":"2026-01-26T15:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.848550 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.854563 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f"} Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.884406 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:16 crc kubenswrapper[4880]: I0126 15:55:16.910802 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:16Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.152154 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.155371 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.155412 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.155424 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.155461 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.155477 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:17Z","lastTransitionTime":"2026-01-26T15:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.213220 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.213284 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.213237 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:17 crc kubenswrapper[4880]: E0126 15:55:17.213427 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:17 crc kubenswrapper[4880]: E0126 15:55:17.213538 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:17 crc kubenswrapper[4880]: E0126 15:55:17.213687 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.227230 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.244603 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.524331 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.527678 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.527733 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.527750 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.527789 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.527819 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:17Z","lastTransitionTime":"2026-01-26T15:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.552038 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.563542 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 06:07:27.744641693 +0000 UTC Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.577551 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.606120 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.630128 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.634467 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.634537 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.634563 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.634601 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.634612 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:17Z","lastTransitionTime":"2026-01-26T15:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.647018 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.661386 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.675547 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:17Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.738094 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.738605 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.738784 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.738916 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:17 crc kubenswrapper[4880]: I0126 15:55:17.739164 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:17Z","lastTransitionTime":"2026-01-26T15:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.121074 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.121119 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.121132 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.121160 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.121173 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:18Z","lastTransitionTime":"2026-01-26T15:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.223037 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.223072 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.223085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.223108 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.223121 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:18Z","lastTransitionTime":"2026-01-26T15:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.327310 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.327353 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.327386 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.327419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.327470 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:18Z","lastTransitionTime":"2026-01-26T15:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.431052 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.431112 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.431124 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.431163 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.431183 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:18Z","lastTransitionTime":"2026-01-26T15:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.910293 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.910356 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.910369 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.910386 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.910398 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:18Z","lastTransitionTime":"2026-01-26T15:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:18 crc kubenswrapper[4880]: I0126 15:55:18.911907 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 22:38:38.625424884 +0000 UTC Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.014862 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.014935 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.014951 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.014991 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.015020 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.118175 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.118573 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.118672 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.118747 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.118849 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.134060 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4" exitCode=0 Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.134237 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.139170 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.162684 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.183368 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.205526 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.213670 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.213809 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:19 crc kubenswrapper[4880]: E0126 15:55:19.213955 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.214486 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:19 crc kubenswrapper[4880]: E0126 15:55:19.214670 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:19 crc kubenswrapper[4880]: E0126 15:55:19.214878 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.224298 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.224406 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.224422 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.224460 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.224475 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.225709 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.241636 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.279794 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.308606 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.329020 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.330706 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.330782 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.330813 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.330846 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.330867 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.357992 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.380068 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.394570 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.407865 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.425624 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.434479 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.434586 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.434623 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.434666 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.434693 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.750614 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.750670 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.750691 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.750719 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.750737 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.757071 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:19Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.854409 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.854556 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.854584 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.854617 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.854631 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.912349 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 01:23:34.790017395 +0000 UTC Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.958471 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.958565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.958597 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.958634 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:19 crc kubenswrapper[4880]: I0126 15:55:19.958664 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:19Z","lastTransitionTime":"2026-01-26T15:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.061624 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.061686 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.061706 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.061726 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.061745 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.149333 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerStarted","Data":"d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.164377 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.164426 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.164458 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.164480 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.164499 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.593589 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.593637 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.593648 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.593666 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.593678 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.594412 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.613931 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.632386 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.736402 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.736494 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.736519 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.736539 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.736552 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.741503 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.767212 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.787628 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.816825 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.840383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.840416 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.840425 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.840463 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.840510 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.843229 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.858273 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.876523 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.893733 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.913663 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:12:43.702142909 +0000 UTC Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.915135 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.936759 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.936897 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:55:36.93668873 +0000 UTC m=+62.422417437 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.937116 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.937201 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.937649 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.937913 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938082 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:36.938038512 +0000 UTC m=+62.423767389 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938251 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938358 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:36.938325818 +0000 UTC m=+62.424054535 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938501 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938523 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938549 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:20 crc kubenswrapper[4880]: E0126 15:55:20.938825 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:36.938793199 +0000 UTC m=+62.424521906 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.939605 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.944943 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.945020 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.945048 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.945076 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.945095 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:20Z","lastTransitionTime":"2026-01-26T15:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:20 crc kubenswrapper[4880]: I0126 15:55:20.990646 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:20Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:21 crc kubenswrapper[4880]: I0126 15:55:21.101398 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:21 crc kubenswrapper[4880]: I0126 15:55:21.101483 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:21 crc kubenswrapper[4880]: I0126 15:55:21.101495 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:21 crc kubenswrapper[4880]: I0126 15:55:21.101520 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:21 crc kubenswrapper[4880]: I0126 15:55:21.101536 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:21Z","lastTransitionTime":"2026-01-26T15:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:21.224042 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:21.224423 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.175733 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 20:25:42.908070304 +0000 UTC Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.177263 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.177380 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.177435 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.177636 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.193570 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.193786 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.193814 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.193831 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.193912 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:38.19389277 +0000 UTC m=+63.679621477 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.203280 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.204835 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.204971 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.205057 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.205140 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.206696 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.206752 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.206767 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.206785 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.206798 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.219844 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f"} Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.224202 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:22Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.230507 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.230551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.230561 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.230577 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.230587 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.246691 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:22Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.252492 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.252815 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.252959 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.253131 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.253223 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.268752 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:22Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.274189 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.274493 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.274600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.274782 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.274896 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.355306 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:22Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.363465 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.363604 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.363613 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.363630 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.363644 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.500037 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:22Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:22 crc kubenswrapper[4880]: E0126 15:55:22.500191 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.517807 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.517869 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.517881 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.517901 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.517913 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.626519 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.626567 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.626579 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.626597 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.626613 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.729546 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.729604 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.729614 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.729630 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.729642 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.984763 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.985176 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.985270 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.985373 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:22 crc kubenswrapper[4880]: I0126 15:55:22.985491 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:22Z","lastTransitionTime":"2026-01-26T15:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.154004 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.154341 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.154494 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.154598 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.154869 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:23Z","lastTransitionTime":"2026-01-26T15:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.176336 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 09:12:37.249625157 +0000 UTC Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.227146 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:23 crc kubenswrapper[4880]: E0126 15:55:23.227334 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.227147 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.257975 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.258028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.258043 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.258067 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.258085 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:23Z","lastTransitionTime":"2026-01-26T15:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.423838 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.424142 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.424171 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.424188 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.424198 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:23Z","lastTransitionTime":"2026-01-26T15:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.569414 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.569502 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.569527 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.569548 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.569561 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:23Z","lastTransitionTime":"2026-01-26T15:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.901831 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.901871 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.901879 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.901894 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:23 crc kubenswrapper[4880]: I0126 15:55:23.901930 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:23Z","lastTransitionTime":"2026-01-26T15:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.008564 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.008620 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.008634 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.008653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.008667 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.111705 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.111796 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.111813 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.111829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.111840 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.177279 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 06:20:35.747109857 +0000 UTC Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.213388 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:24 crc kubenswrapper[4880]: E0126 15:55:24.213868 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.213950 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:24 crc kubenswrapper[4880]: E0126 15:55:24.214175 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.215750 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.215816 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.215830 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.215854 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.215868 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.253571 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.401028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.401122 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.401133 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.401149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.401160 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.503788 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.503839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.503852 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.503868 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.503880 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.607839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.607890 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.607902 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.607921 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.607934 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.747507 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.747889 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.748021 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.748154 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.748263 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.851710 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.851796 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.851809 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.851834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.851847 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.954683 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.954734 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.954744 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.954761 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:24 crc kubenswrapper[4880]: I0126 15:55:24.954771 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:24Z","lastTransitionTime":"2026-01-26T15:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.057892 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.057939 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.057949 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.057965 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.057975 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.167021 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.167068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.167078 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.167094 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.167115 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.177546 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 06:44:18.712605188 +0000 UTC Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.214122 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:25 crc kubenswrapper[4880]: E0126 15:55:25.214521 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.214876 4880 scope.go:117] "RemoveContainer" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.270993 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.271050 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.271062 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.271080 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.271092 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.273299 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6" exitCode=0 Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.273354 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.292186 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.309668 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.337821 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.351049 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.363889 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382795 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382846 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382857 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382874 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382886 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.382925 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.403159 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.423385 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.446622 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.517459 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.536994 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.537104 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.537122 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.537141 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.537153 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.554373 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.581775 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.607586 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.631967 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:25Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.640272 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.640312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.640324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.640342 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.640353 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.744572 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.744629 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.744652 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.744674 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:25 crc kubenswrapper[4880]: I0126 15:55:25.744694 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.847696 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.847748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.847760 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.847779 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.847792 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.950476 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.950538 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.950551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.950571 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:25.950584 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:25Z","lastTransitionTime":"2026-01-26T15:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.054053 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.054127 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.054159 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.054183 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.054197 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.156988 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.157028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.157037 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.157052 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.157104 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.178984 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 16:48:58.592032828 +0000 UTC Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.213561 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.213754 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:26 crc kubenswrapper[4880]: E0126 15:55:26.213874 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:26 crc kubenswrapper[4880]: E0126 15:55:26.213968 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.228208 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.244419 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260497 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260683 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260708 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260718 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260735 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.260744 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.275045 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.290425 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.304578 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.320406 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.334510 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.349755 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.363129 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.363170 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.363182 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.363198 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.363211 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.365928 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.382985 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.404623 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.417005 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.436781 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:26Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.466511 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.466556 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.466569 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.466590 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.466607 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.571649 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.571720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.571739 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.571767 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.571795 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.674816 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.675240 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.675347 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.675501 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.675681 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.779460 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.779889 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.779975 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.780075 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.780155 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.884073 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.884160 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.884175 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.884198 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.884213 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.987367 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.987419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.987464 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.987484 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:26 crc kubenswrapper[4880]: I0126 15:55:26.987495 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:26Z","lastTransitionTime":"2026-01-26T15:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.101183 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.101252 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.101273 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.101301 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.101315 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.184304 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 14:16:06.886707968 +0000 UTC Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.204787 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.204953 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.205111 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.205219 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.205243 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.213431 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:27 crc kubenswrapper[4880]: E0126 15:55:27.213677 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.284850 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06" exitCode=0 Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.284930 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.289357 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.380672 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.388317 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.388383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.388397 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.388415 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.388428 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.419386 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.421312 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.422726 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.447045 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.464608 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.479838 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.492582 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.493398 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.493608 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.493713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.493742 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.500885 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.523827 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.589406 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.597077 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.597115 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.597126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.597143 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.597157 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.614631 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.629224 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.656113 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.685835 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.701148 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.701199 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.701209 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.701226 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.701238 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.731863 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.757480 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.805693 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.805770 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.805801 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.805833 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.805868 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.878500 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.898561 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.909835 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.909899 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.909911 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.909930 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.909943 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:27Z","lastTransitionTime":"2026-01-26T15:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.969237 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:27 crc kubenswrapper[4880]: I0126 15:55:27.987989 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:27Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014487 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014523 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014533 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014560 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.014643 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.034236 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.049769 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.066933 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.104070 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.139680 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.167514 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.203726 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.204636 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 11:33:09.265719059 +0000 UTC Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.209415 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.209507 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.209528 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.209549 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.209564 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.213698 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:28 crc kubenswrapper[4880]: E0126 15:55:28.213832 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.213989 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:28 crc kubenswrapper[4880]: E0126 15:55:28.214170 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.225498 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.253213 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.312681 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.312745 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.312763 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.312784 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.312801 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.417402 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.417484 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.417501 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.417519 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.417545 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.427534 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f" exitCode=0 Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.427623 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.435988 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.439954 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.454536 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.471713 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.486533 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.505055 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.522861 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.522900 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.522912 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.522930 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.523696 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.525020 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.542757 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.560072 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.584281 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.609011 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.631071 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.631139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.631166 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.631197 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.631210 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.637761 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.665945 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.695427 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.712554 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.734519 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.734595 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.734608 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.734637 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.734651 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.741658 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.758655 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.783215 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.807686 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.827340 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.843861 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.858282 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.876105 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.876149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.876163 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.876182 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.876196 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.877740 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.896580 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.912738 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.932012 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.959877 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.980536 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.980633 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.980647 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.980663 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.980692 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:28Z","lastTransitionTime":"2026-01-26T15:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:28 crc kubenswrapper[4880]: I0126 15:55:28.996941 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:28Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.015788 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.045275 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.084279 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.084326 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.084337 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.084355 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.084368 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.188411 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.188496 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.188510 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.188527 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.188540 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.204834 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 17:55:03.157584352 +0000 UTC Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.213091 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:29 crc kubenswrapper[4880]: E0126 15:55:29.213297 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.292830 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.292897 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.292911 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.292930 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.292943 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.395349 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.395396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.395467 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.395499 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.395513 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.460154 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerStarted","Data":"5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.483392 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.499254 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.499312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.499324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.499343 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.499357 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.500106 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.518565 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.568235 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.585882 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.601745 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.601805 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.601815 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.601832 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.601845 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.648568 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.664275 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.681422 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.695652 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.704915 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.705028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.705046 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.705073 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.705092 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.708590 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.721945 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.742598 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.762756 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.781125 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:29Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.820984 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.821076 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.821097 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.821121 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.821136 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.924255 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.924306 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.924321 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.924342 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:29 crc kubenswrapper[4880]: I0126 15:55:29.924359 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:29Z","lastTransitionTime":"2026-01-26T15:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.057011 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.057720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.057756 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.057795 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.057815 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.160645 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.160695 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.160705 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.160726 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.160738 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.205592 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 20:11:25.211661945 +0000 UTC Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.214338 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.214424 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:30 crc kubenswrapper[4880]: E0126 15:55:30.214639 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:30 crc kubenswrapper[4880]: E0126 15:55:30.214743 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.264350 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.264392 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.264404 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.264422 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.264450 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.459622 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.459703 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.459719 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.459739 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.459751 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.536407 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.538818 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.539122 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.539150 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.560634 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.568161 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.568256 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.568281 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.568322 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.568359 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.600211 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.630083 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.663237 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.673851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.673925 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.673944 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.673983 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.674020 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.700123 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.724765 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.744470 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.764133 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.771376 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.779062 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.779118 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.779130 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.779152 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.779167 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.882628 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.882675 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.882688 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.882711 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.882729 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:30 crc kubenswrapper[4880]: I0126 15:55:30.999907 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:30.999964 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:30.999977 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:30.999998 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.000012 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:30Z","lastTransitionTime":"2026-01-26T15:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.003860 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:30Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.012792 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.025130 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.186282 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.186382 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.186396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.186417 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.186452 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.193796 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.205932 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 03:45:21.158412547 +0000 UTC Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.213720 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:31 crc kubenswrapper[4880]: E0126 15:55:31.213919 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.219052 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.241587 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.259898 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.279242 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.289369 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.289410 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.289423 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.289459 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.289474 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.295523 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.313376 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.339292 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.356681 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.386615 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.406968 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.435289 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.453843 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.475320 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.501173 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.523859 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.542769 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.555718 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.555826 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.555842 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.555864 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.555876 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.565635 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.661021 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.661078 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.661091 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.661111 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.661123 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.764052 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.764137 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.764148 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.764167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.764185 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.872480 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.872551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.872565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.872594 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.872610 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.937997 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz"] Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.938675 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.941848 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.942534 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.953817 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.972524 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.975265 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.975308 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.975319 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.975336 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.975348 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:31Z","lastTransitionTime":"2026-01-26T15:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:31 crc kubenswrapper[4880]: I0126 15:55:31.992166 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:31Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.032346 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.049601 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.070149 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.078220 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.078558 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.078675 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.078856 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.078991 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.089101 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.093345 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2z26\" (UniqueName: \"kubernetes.io/projected/5beb13dc-5f14-4757-b9a0-01d074dfbd23-kube-api-access-x2z26\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.094684 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.094935 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.095064 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.110378 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.129392 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.149525 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.185990 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.186423 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.186713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.186957 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.187037 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.196795 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.198533 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2z26\" (UniqueName: \"kubernetes.io/projected/5beb13dc-5f14-4757-b9a0-01d074dfbd23-kube-api-access-x2z26\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.199013 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.201334 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.201235 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.203156 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.205662 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5beb13dc-5f14-4757-b9a0-01d074dfbd23-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.206417 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 14:23:11.647704989 +0000 UTC Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.213405 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.214009 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.215469 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.215854 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.228160 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5beb13dc-5f14-4757-b9a0-01d074dfbd23-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.229839 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.233365 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2z26\" (UniqueName: \"kubernetes.io/projected/5beb13dc-5f14-4757-b9a0-01d074dfbd23-kube-api-access-x2z26\") pod \"ovnkube-control-plane-749d76644c-jrmsz\" (UID: \"5beb13dc-5f14-4757-b9a0-01d074dfbd23\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.256702 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.269747 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290027 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290132 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290147 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290173 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290189 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.290627 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.313862 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.393768 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.393818 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.393841 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.393861 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.393875 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.412517 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.429044 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.432883 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.460173 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.479950 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.496945 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.496989 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.497000 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.497017 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.497029 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.497563 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.513905 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.530189 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.547027 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.569864 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.575289 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" event={"ID":"5beb13dc-5f14-4757-b9a0-01d074dfbd23","Type":"ContainerStarted","Data":"b0563e3aace7364861ef5faf2f75321b2d2122d140972d2a5532e16980308741"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.586848 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.592684 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.592737 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.592748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.592768 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.592782 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.606609 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.609340 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.615784 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.615829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.615841 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.615857 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.615868 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.622801 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.631993 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.637580 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.637650 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.637661 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.637680 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.637692 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.640174 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.650651 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.653789 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.658902 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.658949 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.658960 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.658978 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.658990 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.691637 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.700858 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.707500 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.707542 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.707553 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.707572 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.707584 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.708655 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.737893 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:32Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:32 crc kubenswrapper[4880]: E0126 15:55:32.738041 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.740729 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.740780 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.740791 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.740810 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.740821 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.844161 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.844205 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.844218 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.844240 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.844253 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.949667 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.949846 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.951358 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.951630 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:32 crc kubenswrapper[4880]: I0126 15:55:32.951649 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:32Z","lastTransitionTime":"2026-01-26T15:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.057019 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.057478 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.057700 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.057853 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.057993 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.080481 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-pvdts"] Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.083854 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.083972 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.101716 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.117368 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.117371 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.117460 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgp6d\" (UniqueName: \"kubernetes.io/projected/216872c9-037e-486f-bf08-5e970a2bd5b1-kube-api-access-xgp6d\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.134220 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.147135 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.162187 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.162252 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.162266 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.162285 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.162302 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.163267 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.177262 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.189742 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.200696 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.207690 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 05:34:10.721019911 +0000 UTC Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.212320 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.213539 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.213721 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.218915 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.218977 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgp6d\" (UniqueName: \"kubernetes.io/projected/216872c9-037e-486f-bf08-5e970a2bd5b1-kube-api-access-xgp6d\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.219262 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.219322 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:33.719303956 +0000 UTC m=+59.205032663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.225821 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.239055 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgp6d\" (UniqueName: \"kubernetes.io/projected/216872c9-037e-486f-bf08-5e970a2bd5b1-kube-api-access-xgp6d\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.239779 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.255704 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.265601 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.265643 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.265652 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.265688 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.265698 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.271545 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.329418 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.355786 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370196 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370208 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370227 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370240 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.370903 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.383498 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.481653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.481733 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.481746 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.481764 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.481774 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.583634 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c" exitCode=0 Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.583700 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.584064 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.584172 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.584186 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.584207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.584228 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.587198 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" event={"ID":"5beb13dc-5f14-4757-b9a0-01d074dfbd23","Type":"ContainerStarted","Data":"8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.603936 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.620895 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.639922 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.653730 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.672492 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.687176 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.687222 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.687232 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.687247 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.687256 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.688141 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.702721 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.718922 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.723546 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.723730 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:33 crc kubenswrapper[4880]: E0126 15:55:33.723800 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:34.723781226 +0000 UTC m=+60.209509933 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.788036 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.791489 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.791534 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.791546 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.791563 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.791576 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.808277 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.832040 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.848601 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.862454 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.877361 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.893494 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.895370 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.895432 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.895463 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.895512 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.895532 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.908601 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.924603 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.941490 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.957215 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.972379 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.989086 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:33Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.999080 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.999156 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.999172 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.999198 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:33 crc kubenswrapper[4880]: I0126 15:55:33.999243 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:33Z","lastTransitionTime":"2026-01-26T15:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.008241 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.033881 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.053331 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.075400 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.091528 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.102174 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.102225 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.102243 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.102263 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.102276 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.113725 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.126853 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.139737 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.156324 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.178918 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.292896 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 09:39:30.705710098 +0000 UTC Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.293012 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.293094 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:34 crc kubenswrapper[4880]: E0126 15:55:34.293320 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:34 crc kubenswrapper[4880]: E0126 15:55:34.293468 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.295073 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.295106 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.295118 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.295135 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.295158 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.296389 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.315418 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.333551 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.412190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.412269 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.412300 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.412347 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.412366 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.516524 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.516580 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.516593 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.516615 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.516629 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.595774 4880 generic.go:334] "Generic (PLEG): container finished" podID="8dbe43e2-b41e-4bbe-8bb3-c3389816a067" containerID="186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150" exitCode=0 Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.595901 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerDied","Data":"186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.599191 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" event={"ID":"5beb13dc-5f14-4757-b9a0-01d074dfbd23","Type":"ContainerStarted","Data":"3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.620919 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.620992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.621005 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.621023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.621034 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.624896 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.641276 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.656098 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.671354 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.688551 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.703675 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.715156 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.729305 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.743696 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.758044 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.771796 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:34 crc kubenswrapper[4880]: E0126 15:55:34.771964 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:34 crc kubenswrapper[4880]: E0126 15:55:34.772057 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:36.772035947 +0000 UTC m=+62.257764654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.774055 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.774098 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.774107 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.774122 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.774133 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.778956 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.795765 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.809561 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.829940 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.844226 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.859794 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.877171 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.877224 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.877237 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.877256 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.877268 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.887554 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:34Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.981457 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.981516 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.981541 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.981569 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:34 crc kubenswrapper[4880]: I0126 15:55:34.981592 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:34Z","lastTransitionTime":"2026-01-26T15:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.085358 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.085418 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.085446 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.085466 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.085475 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.188339 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.188383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.188397 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.188414 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.188424 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.213977 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.214032 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:35 crc kubenswrapper[4880]: E0126 15:55:35.214227 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:35 crc kubenswrapper[4880]: E0126 15:55:35.214326 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.291669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.291736 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.291748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.291772 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.291786 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.293831 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 23:26:35.017222857 +0000 UTC Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.396588 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.396643 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.396654 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.396671 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.396683 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.499840 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.499886 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.499898 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.499915 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.499927 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.602126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.602197 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.602209 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.602229 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.602242 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.607901 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" event={"ID":"8dbe43e2-b41e-4bbe-8bb3-c3389816a067","Type":"ContainerStarted","Data":"d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.620898 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.636941 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.662311 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.678091 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.700325 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.705757 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.705843 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.705867 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.705885 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.705896 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.725516 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.750428 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.768855 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.793526 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.810718 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.811141 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.811274 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.811400 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.811547 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.816225 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.859555 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.994141 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.994276 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.994301 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.994325 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:35 crc kubenswrapper[4880]: I0126 15:55:35.994343 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:35Z","lastTransitionTime":"2026-01-26T15:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.008051 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:35Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.027400 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.110734 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.112736 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.112935 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.113015 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.113110 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.113227 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.132525 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.150030 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.165056 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.213718 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.213853 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:36 crc kubenswrapper[4880]: E0126 15:55:36.213885 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:36 crc kubenswrapper[4880]: E0126 15:55:36.214028 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.216657 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.216691 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.216703 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.216718 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.216730 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.227329 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.247865 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.267648 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.284399 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.294378 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 23:53:07.072856971 +0000 UTC Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.301675 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.320866 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.320937 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.320951 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.320969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.320981 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.325848 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.344784 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.364166 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.379219 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.402956 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.417113 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.423616 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.423659 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.423669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.423684 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.423692 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.434733 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.458326 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.526831 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.526910 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.526923 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.526942 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.526954 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.629909 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.629944 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.629954 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.629969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.629980 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.715358 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.737146 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.737220 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.737235 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.737259 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.737278 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.751031 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.769903 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.790382 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:36Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.822409 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:36 crc kubenswrapper[4880]: E0126 15:55:36.822639 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:36 crc kubenswrapper[4880]: E0126 15:55:36.822732 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:40.822707063 +0000 UTC m=+66.308435770 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.840389 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.840449 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.840460 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.840473 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.840483 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.943659 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.943702 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.943724 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.943752 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:36 crc kubenswrapper[4880]: I0126 15:55:36.943779 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:36Z","lastTransitionTime":"2026-01-26T15:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.024715 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.024942 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025288 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:09.025253841 +0000 UTC m=+94.510982548 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.025364 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.025490 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025549 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025596 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025614 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025646 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025676 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:09.025658881 +0000 UTC m=+94.511387588 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025568 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025744 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:09.025734543 +0000 UTC m=+94.511463250 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.025798 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:09.025777824 +0000 UTC m=+94.511506531 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.048965 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.049048 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.049068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.049102 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.049149 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.153713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.153766 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.153786 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.153812 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.153832 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.213481 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.213603 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.213677 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:37 crc kubenswrapper[4880]: E0126 15:55:37.213751 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.263770 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.263820 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.263830 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.263849 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.263862 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.295186 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 01:39:10.443741987 +0000 UTC Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.367322 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.367427 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.367488 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.367514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.367582 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.470911 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.470978 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.470990 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.471014 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.471028 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.577631 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.577687 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.577700 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.577719 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.577734 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.681565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.681654 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.681668 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.681714 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.681733 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.785639 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.785697 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.785709 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.785728 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.785741 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.890703 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.890748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.890763 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.890781 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.890793 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.995165 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.995829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.995956 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.996057 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:37 crc kubenswrapper[4880]: I0126 15:55:37.996157 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:37Z","lastTransitionTime":"2026-01-26T15:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.099754 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.099806 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.099818 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.099837 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.099851 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.213743 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.214159 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.215675 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.215777 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.239554 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.239818 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.239864 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.239882 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:38 crc kubenswrapper[4880]: E0126 15:55:38.239971 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:10.239943844 +0000 UTC m=+95.725672551 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.248501 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.248556 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.248568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.248586 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.248597 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.328796 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 02:37:43.978089848 +0000 UTC Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.351426 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.351530 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.351542 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.351561 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.351580 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.454097 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.454154 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.454166 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.454183 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.454206 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.557910 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.557969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.557982 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.558001 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.558015 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.660626 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.660669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.660682 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.660701 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.660714 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.763945 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.763995 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.764005 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.764022 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.764033 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.866221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.866268 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.866285 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.866306 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.866319 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.969724 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.969770 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.969790 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.969806 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:38 crc kubenswrapper[4880]: I0126 15:55:38.969816 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:38Z","lastTransitionTime":"2026-01-26T15:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.077102 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.077161 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.077173 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.077193 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.077205 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.180009 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.180066 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.180085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.180103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.180114 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.213410 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.213495 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:39 crc kubenswrapper[4880]: E0126 15:55:39.213618 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:39 crc kubenswrapper[4880]: E0126 15:55:39.213715 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.283057 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.283117 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.283131 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.283151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.283166 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.329636 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 01:21:03.881879147 +0000 UTC Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.386551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.386612 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.386624 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.386644 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.386656 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.489397 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.489476 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.489490 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.489504 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.489514 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.592371 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.592478 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.592489 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.592511 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.592548 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.627747 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/0.log" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.633234 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.634091 4880 scope.go:117] "RemoveContainer" containerID="d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.634991 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465" exitCode=1 Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.661032 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.679884 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.695761 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.698051 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.698102 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.698118 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.698136 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.698149 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.720579 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.735911 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.750421 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.772656 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:39Z\\\",\\\"message\\\":\\\"wall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.453940 6116 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454097 6116 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 15:55:38.454229 6116 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454549 6116 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:38.454572 6116 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:38.454623 6116 factory.go:656] Stopping watch factory\\\\nI0126 15:55:38.454642 6116 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:38.454651 6116 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:38.454677 6116 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.787278 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.800813 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.800886 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.800905 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.800923 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.800973 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.804956 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.823413 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.852265 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.877794 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.896700 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.907068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.907123 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.907151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.907181 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.907210 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:39Z","lastTransitionTime":"2026-01-26T15:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.913074 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.933090 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.954227 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:39 crc kubenswrapper[4880]: I0126 15:55:39.978118 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:39Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.012465 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.012528 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.012552 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.012580 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.012605 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.120140 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.120188 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.120199 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.120215 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.120226 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.217062 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:40 crc kubenswrapper[4880]: E0126 15:55:40.217590 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.217860 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:40 crc kubenswrapper[4880]: E0126 15:55:40.217927 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.223724 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.223773 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.223787 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.223803 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.223815 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.327495 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.327542 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.327986 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.328021 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.328043 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.337727 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 12:37:42.480847523 +0000 UTC Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.432017 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.432095 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.432108 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.432137 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.432152 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.536082 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.536147 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.536160 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.536193 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.536210 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.639795 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.639832 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.639841 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.639858 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.639869 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.642701 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/0.log" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.644918 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.646002 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.668949 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:39Z\\\",\\\"message\\\":\\\"wall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.453940 6116 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454097 6116 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 15:55:38.454229 6116 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454549 6116 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:38.454572 6116 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:38.454623 6116 factory.go:656] Stopping watch factory\\\\nI0126 15:55:38.454642 6116 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:38.454651 6116 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:38.454677 6116 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.683646 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.697702 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.711185 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.728733 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.743330 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.743374 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.743388 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.743407 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.743419 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.746091 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.761994 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.776103 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.793488 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.812266 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.830472 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.843726 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.846135 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.846169 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.846180 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.846194 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.846203 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.858610 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.870699 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:40 crc kubenswrapper[4880]: E0126 15:55:40.870903 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:40 crc kubenswrapper[4880]: E0126 15:55:40.870971 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:55:48.870950921 +0000 UTC m=+74.356679638 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.874055 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.888934 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.911625 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.928110 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:40Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.948980 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.949023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.949035 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.949051 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:40 crc kubenswrapper[4880]: I0126 15:55:40.949062 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:40Z","lastTransitionTime":"2026-01-26T15:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.052206 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.052246 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.052258 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.052276 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.052289 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.154992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.155038 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.155049 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.155068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.155083 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.213630 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:41 crc kubenswrapper[4880]: E0126 15:55:41.213783 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.214171 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:41 crc kubenswrapper[4880]: E0126 15:55:41.214253 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.424348 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 06:31:38.949541986 +0000 UTC Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.426637 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.426685 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.426703 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.426730 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.426742 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.531812 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.531941 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.531977 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.532051 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.532085 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.636677 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.637316 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.637623 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.637784 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.637890 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.837743 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.837819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.837829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.837847 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.837857 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.940661 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.940708 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.940716 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.940731 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:41 crc kubenswrapper[4880]: I0126 15:55:41.940749 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:41Z","lastTransitionTime":"2026-01-26T15:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.043545 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.044149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.044337 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.044505 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.044640 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.147245 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.147310 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.147333 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.147353 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.147367 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.213555 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.213579 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:42 crc kubenswrapper[4880]: E0126 15:55:42.213729 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:42 crc kubenswrapper[4880]: E0126 15:55:42.213918 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.250846 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.250901 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.250928 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.250947 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.250964 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.354113 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.354172 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.354188 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.354211 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.354234 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.424901 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 15:53:08.635237891 +0000 UTC Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.458353 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.458429 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.458467 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.458508 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.458521 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.562039 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.562125 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.562140 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.562167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.562186 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.655834 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/1.log" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.656872 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/0.log" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.660143 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf" exitCode=1 Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.660198 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.660241 4880 scope.go:117] "RemoveContainer" containerID="d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.661234 4880 scope.go:117] "RemoveContainer" containerID="9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf" Jan 26 15:55:42 crc kubenswrapper[4880]: E0126 15:55:42.661469 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2p7jp_openshift-ovn-kubernetes(c377ad66-a182-4cbf-9b29-253f3e547118)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.665492 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.665565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.665582 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.665601 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.665615 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.674421 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.696860 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:39Z\\\",\\\"message\\\":\\\"wall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.453940 6116 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454097 6116 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 15:55:38.454229 6116 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454549 6116 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:38.454572 6116 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:38.454623 6116 factory.go:656] Stopping watch factory\\\\nI0126 15:55:38.454642 6116 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:38.454651 6116 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:38.454677 6116 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.711143 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.727022 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.752828 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.769654 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.771345 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.771558 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.771656 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.771765 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.771865 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.786879 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.809847 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.830044 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.847659 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.875335 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.882550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.882600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.882653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.882695 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.882710 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.892133 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.908999 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.928527 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.948736 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.968278 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.987028 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.987119 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.987130 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.987181 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.987202 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:42Z","lastTransitionTime":"2026-01-26T15:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:42 crc kubenswrapper[4880]: I0126 15:55:42.990987 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:42Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.091574 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.091643 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.091654 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.091687 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.091703 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.118786 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.118852 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.118866 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.118889 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.118987 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.136761 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:43Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.141861 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.141906 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.141914 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.141931 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.141945 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.156115 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:43Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.160890 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.160934 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.160945 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.160963 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.160976 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.176034 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:43Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.184956 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.185032 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.185046 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.185068 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.185081 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.200616 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:43Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.204819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.204944 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.204954 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.204969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.204980 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.214159 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.214325 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.214414 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.214544 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.219733 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:43Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:43 crc kubenswrapper[4880]: E0126 15:55:43.219866 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.223091 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.223131 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.223141 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.223161 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.223175 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.325735 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.325884 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.326271 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.326628 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.326907 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.425694 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 04:42:18.631001224 +0000 UTC Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.429981 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.430020 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.430030 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.430047 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.430063 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.533570 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.533627 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.533637 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.533657 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.533670 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.638045 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.638536 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.638672 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.638797 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.638914 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.672340 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/1.log" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.743581 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.743641 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.743655 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.743679 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.743695 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.847809 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.848815 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.848978 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.849095 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.849180 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.955902 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.955958 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.955969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.955988 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:43 crc kubenswrapper[4880]: I0126 15:55:43.956004 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:43Z","lastTransitionTime":"2026-01-26T15:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.061649 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.061718 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.061734 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.061783 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.061797 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.165244 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.165293 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.165306 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.165326 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.165339 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.213877 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.213949 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:44 crc kubenswrapper[4880]: E0126 15:55:44.214112 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:44 crc kubenswrapper[4880]: E0126 15:55:44.214287 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.268495 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.268587 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.268834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.269172 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.269230 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.328670 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.348789 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.367219 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.374074 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.374126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.374149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.374167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.374180 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.383643 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.400236 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.415182 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.426483 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 14:31:32.19051465 +0000 UTC Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.435179 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.454659 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.473706 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.477030 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.477096 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.477116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.477139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.477153 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.495401 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.510604 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.533768 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:39Z\\\",\\\"message\\\":\\\"wall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.453940 6116 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454097 6116 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 15:55:38.454229 6116 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454549 6116 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:38.454572 6116 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:38.454623 6116 factory.go:656] Stopping watch factory\\\\nI0126 15:55:38.454642 6116 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:38.454651 6116 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:38.454677 6116 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.557516 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.580639 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.583083 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.583143 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.583159 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.583185 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.583198 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.610734 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.764079 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.764137 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.764149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.764167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.764182 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.770568 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.791678 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.810308 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:44Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.866999 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.867041 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.867051 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.867065 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.867076 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.969642 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.969699 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.969711 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.969730 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:44 crc kubenswrapper[4880]: I0126 15:55:44.969754 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:44Z","lastTransitionTime":"2026-01-26T15:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.072969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.073022 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.073033 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.073049 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.073060 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.177051 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.177103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.177113 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.177132 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.177144 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.213530 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.213564 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:45 crc kubenswrapper[4880]: E0126 15:55:45.213720 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:45 crc kubenswrapper[4880]: E0126 15:55:45.213837 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.281245 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.281295 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.281305 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.281325 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.281335 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.384514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.384572 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.384585 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.384605 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.384616 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.427592 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 01:27:30.661422394 +0000 UTC Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.488273 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.488380 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.488395 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.488421 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.488665 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.591822 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.591903 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.591917 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.591937 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.591956 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.695887 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.695940 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.695952 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.695969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.695981 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.799736 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.799778 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.799790 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.799808 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:45 crc kubenswrapper[4880]: I0126 15:55:45.799820 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:45Z","lastTransitionTime":"2026-01-26T15:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.017283 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.017347 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.017367 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.017389 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.017404 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.121376 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.121447 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.121459 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.121477 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.121492 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.213935 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.214000 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:46 crc kubenswrapper[4880]: E0126 15:55:46.214279 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:46 crc kubenswrapper[4880]: E0126 15:55:46.214390 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.225822 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.225875 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.225885 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.225903 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.225917 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.232589 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.259710 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75d2fc1423db1899eaa743a9b7974f7fd48b30104aa0ce4e7d32e14ce98a465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:39Z\\\",\\\"message\\\":\\\"wall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.453940 6116 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454097 6116 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 15:55:38.454229 6116 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 15:55:38.454549 6116 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:38.454572 6116 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:38.454623 6116 factory.go:656] Stopping watch factory\\\\nI0126 15:55:38.454642 6116 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:38.454651 6116 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:38.454677 6116 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.279105 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.298394 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.314725 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.329036 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.329087 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.329098 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.329114 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.329128 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.334725 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.354229 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.373553 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.392647 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.430943 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:04:56.659671868 +0000 UTC Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.582095 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.586505 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.586550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.586560 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.586583 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.586597 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.607861 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.625396 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.641370 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.660286 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.680010 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.690182 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.690531 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.690642 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.690746 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.690819 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.698355 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.719351 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:46Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.793004 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.793693 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.793833 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.793957 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.794083 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.897499 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.897568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.897583 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.897606 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:46 crc kubenswrapper[4880]: I0126 15:55:46.897621 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:46Z","lastTransitionTime":"2026-01-26T15:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.001064 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.001100 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.001110 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.001176 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.001205 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.104302 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.104525 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.104541 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.104559 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.104572 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.209014 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213130 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213290 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213223 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213578 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: E0126 15:55:47.213596 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213223 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.213769 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: E0126 15:55:47.214064 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.317027 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.317402 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.317585 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.318678 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.318857 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.422288 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.422338 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.422349 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.422373 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.422385 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.434612 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 18:58:51.322920213 +0000 UTC Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.527150 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.527224 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.527264 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.527316 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.527339 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.630494 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.630548 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.630559 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.630584 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.630597 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.745827 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.745894 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.746209 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.746234 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.746252 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.849835 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.849889 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.849900 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.849917 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.849927 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.953496 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.953906 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.954015 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.954151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:47 crc kubenswrapper[4880]: I0126 15:55:47.954256 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:47Z","lastTransitionTime":"2026-01-26T15:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.058493 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.058565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.058586 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.058612 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.058626 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.161998 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.162081 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.162099 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.162119 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.162132 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.214267 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.214284 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:48 crc kubenswrapper[4880]: E0126 15:55:48.214520 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:48 crc kubenswrapper[4880]: E0126 15:55:48.214631 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.264778 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.264839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.264850 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.264867 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.264880 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.368252 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.368372 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.368385 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.368404 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.368416 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.435018 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 13:15:45.488888231 +0000 UTC Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.471606 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.471664 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.471677 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.471695 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.471706 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.575294 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.575370 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.575382 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.575450 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.575463 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.850897 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.850937 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.850948 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.850970 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.850984 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.949148 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:48 crc kubenswrapper[4880]: E0126 15:55:48.949544 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:48 crc kubenswrapper[4880]: E0126 15:55:48.949706 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:04.949677968 +0000 UTC m=+90.435406715 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.955063 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.955099 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.955109 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.955124 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:48 crc kubenswrapper[4880]: I0126 15:55:48.955134 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:48Z","lastTransitionTime":"2026-01-26T15:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.057385 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.057474 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.057491 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.057514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.057529 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.160751 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.161164 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.161242 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.161327 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.161401 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.213874 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.213975 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:49 crc kubenswrapper[4880]: E0126 15:55:49.214880 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:49 crc kubenswrapper[4880]: E0126 15:55:49.214908 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.264939 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.265077 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.265111 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.265130 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.265148 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.368664 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.368728 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.368741 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.368761 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.368776 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.435473 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 22:47:49.275471104 +0000 UTC Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.472157 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.472217 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.472234 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.472254 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.472270 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.575362 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.575410 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.575419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.575448 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.575459 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.677801 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.677839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.677848 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.677862 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.677872 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.843578 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.843632 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.843643 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.843664 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.843677 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.946641 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.946700 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.946720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.946744 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:49 crc kubenswrapper[4880]: I0126 15:55:49.946778 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:49Z","lastTransitionTime":"2026-01-26T15:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.049391 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.049483 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.049498 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.049521 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.049534 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.152916 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.152979 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.152992 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.153010 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.153020 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.214067 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.214148 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:50 crc kubenswrapper[4880]: E0126 15:55:50.214318 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:50 crc kubenswrapper[4880]: E0126 15:55:50.214497 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.256355 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.256419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.256646 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.256666 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.256677 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.359713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.359767 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.359783 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.359802 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.359822 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.436334 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 11:58:30.565061839 +0000 UTC Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.463256 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.463307 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.463320 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.463337 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.463359 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.566289 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.566345 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.566362 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.566381 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.566391 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.669211 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.669285 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.669306 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.669327 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.669339 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.772324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.772377 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.772390 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.772409 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.772422 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.875966 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.876012 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.876022 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.903969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:50 crc kubenswrapper[4880]: I0126 15:55:50.945977 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:50Z","lastTransitionTime":"2026-01-26T15:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.051829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.051867 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.051877 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.051894 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.051910 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.155221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.155276 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.155289 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.155306 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.155318 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.214040 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.214370 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:51 crc kubenswrapper[4880]: E0126 15:55:51.214544 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:51 crc kubenswrapper[4880]: E0126 15:55:51.214854 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.258364 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.258425 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.258461 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.258484 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.258497 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.361748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.361800 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.361812 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.361831 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.361841 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.437119 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 07:05:18.99126513 +0000 UTC Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.465064 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.465120 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.465135 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.465157 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.465173 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.568618 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.568670 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.568680 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.568695 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.568708 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.672498 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.672564 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.672578 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.672599 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.672612 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.775260 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.775324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.775339 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.775359 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.775375 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.879005 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.879061 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.879074 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.879092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.879104 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.983132 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.983191 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.983204 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.983222 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:51 crc kubenswrapper[4880]: I0126 15:55:51.983236 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:51Z","lastTransitionTime":"2026-01-26T15:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.086226 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.086283 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.086294 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.086313 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.086324 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.189514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.189568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.189577 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.189594 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.189605 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.213048 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:52 crc kubenswrapper[4880]: E0126 15:55:52.213189 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.213414 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:52 crc kubenswrapper[4880]: E0126 15:55:52.213585 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.292493 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.292547 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.292568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.292590 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.292609 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.395860 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.395922 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.395936 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.395990 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.396008 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.438064 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:04:54.302842417 +0000 UTC Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.499330 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.499386 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.499396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.499414 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.499424 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.602324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.602485 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.602500 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.602517 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.602527 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.705000 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.705040 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.705048 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.705063 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.705073 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.808055 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.808092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.808101 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.808116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.808126 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.910593 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.910637 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.910649 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.910667 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:52 crc kubenswrapper[4880]: I0126 15:55:52.910681 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:52Z","lastTransitionTime":"2026-01-26T15:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.014536 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.014600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.014613 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.014633 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.014645 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.117943 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.117993 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.118005 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.118022 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.118037 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.213675 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.213855 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.213914 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.213980 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.221372 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.221463 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.221478 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.221500 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.221514 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.325570 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.325618 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.325635 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.325653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.325668 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.434808 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.434875 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.434889 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.434909 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.434921 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.438466 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 01:25:20.849981995 +0000 UTC Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.536626 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.536721 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.536732 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.536753 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.536810 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.555670 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:53Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.561023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.561103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.561126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.561147 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.561159 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.580003 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:53Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.585469 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.585520 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.585534 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.585555 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.585567 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.606087 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:53Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.611639 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.611704 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.611721 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.611743 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.611758 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.628639 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:53Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.636783 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.636858 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.636877 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.636899 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.636912 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.654699 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:53Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:53 crc kubenswrapper[4880]: E0126 15:55:53.654839 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.656851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.656902 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.656915 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.656940 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.656964 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.760419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.760492 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.760504 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.760524 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.760540 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.864796 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.865242 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.865338 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.865521 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.865614 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.968247 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.968300 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.968311 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.968328 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:53 crc kubenswrapper[4880]: I0126 15:55:53.968339 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:53Z","lastTransitionTime":"2026-01-26T15:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.071584 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.071661 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.071674 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.071695 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.071707 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.175221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.175283 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.175295 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.175312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.175324 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.213897 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:54 crc kubenswrapper[4880]: E0126 15:55:54.214100 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.214302 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:54 crc kubenswrapper[4880]: E0126 15:55:54.215241 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.215710 4880 scope.go:117] "RemoveContainer" containerID="9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.234323 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.265626 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2p7jp_openshift-ovn-kubernetes(c377ad66-a182-4cbf-9b29-253f3e547118)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.281180 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.281227 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.281238 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.281255 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.281267 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.286317 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.303409 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.325211 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.350202 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.375726 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.386982 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.387139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.387158 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.387227 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.387246 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.401408 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.419663 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.439482 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 13:11:08.919239733 +0000 UTC Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.441048 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.462421 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.484832 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.490319 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.490368 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.490381 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.490398 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.490409 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.507739 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.528673 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.553268 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.575796 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.594706 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.594774 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.594792 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.594814 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.594828 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.598540 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.699302 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.699355 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.699370 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.699389 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.699406 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.802989 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.803040 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.803052 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.803069 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.803081 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.881574 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/1.log" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.884823 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.885397 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.906251 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.906300 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.906312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.906338 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.906352 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:54Z","lastTransitionTime":"2026-01-26T15:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.908936 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.934478 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.968691 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:54 crc kubenswrapper[4880]: I0126 15:55:54.997423 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:54Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.012474 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.012525 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.012540 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.012558 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.012570 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.020782 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.045009 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.065378 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.083706 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.105824 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.115852 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.115918 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.115932 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.115959 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.115975 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.128094 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.150781 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.172399 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.194448 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.210153 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.213738 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:55 crc kubenswrapper[4880]: E0126 15:55:55.213934 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.213746 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:55 crc kubenswrapper[4880]: E0126 15:55:55.214251 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.219383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.219487 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.219508 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.219535 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.219552 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.265643 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.284517 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.300179 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:55Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.323514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.323892 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.323995 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.324092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.324167 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.428027 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.428829 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.429151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.429491 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.429620 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.440382 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 12:37:51.213014656 +0000 UTC Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.533996 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.534077 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.534092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.534116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.534134 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.637170 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.637539 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.637646 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.637743 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.637878 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.745412 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.745689 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.745699 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.745714 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.745725 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.848246 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.848688 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.848834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.848955 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.849065 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.952497 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.952563 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.952575 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.952600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:55 crc kubenswrapper[4880]: I0126 15:55:55.952649 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:55Z","lastTransitionTime":"2026-01-26T15:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.055690 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.055742 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.055754 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.055772 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.055784 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.158745 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.158810 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.158819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.158845 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.158856 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.214392 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:56 crc kubenswrapper[4880]: E0126 15:55:56.214605 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.214858 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:56 crc kubenswrapper[4880]: E0126 15:55:56.215135 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.233720 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.250627 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.262529 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.262584 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.262596 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.262616 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.262628 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.265971 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.283752 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.300711 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.318488 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.334307 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.351022 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.365815 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.365870 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.365883 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.365900 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.365912 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.372852 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.390801 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.416639 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.432414 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.440609 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 00:13:57.952607022 +0000 UTC Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.446614 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.459478 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.468953 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.469004 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.469018 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.469037 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.469051 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.472573 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.484037 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.504367 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.572148 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.572254 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.572272 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.572296 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.572307 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.675392 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.675459 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.675470 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.675486 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.675495 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.778157 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.778214 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.778228 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.778248 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.778261 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.881002 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.881374 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.881491 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.881602 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.881805 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.897534 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/2.log" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.898477 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/1.log" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.902769 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" exitCode=1 Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.902828 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.903132 4880 scope.go:117] "RemoveContainer" containerID="9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.904059 4880 scope.go:117] "RemoveContainer" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 15:55:56 crc kubenswrapper[4880]: E0126 15:55:56.904347 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2p7jp_openshift-ovn-kubernetes(c377ad66-a182-4cbf-9b29-253f3e547118)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.924745 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.939347 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.961516 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.976335 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.984447 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.984493 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.984509 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.984525 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.984536 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:56Z","lastTransitionTime":"2026-01-26T15:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:56 crc kubenswrapper[4880]: I0126 15:55:56.992343 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:56Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.007460 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.020676 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.034899 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.051128 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.062920 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.088059 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.088139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.088170 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.088207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.088244 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.092946 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:55Z\\\",\\\"message\\\":\\\"5:55:55.676520 6659 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 15:55:55.678224 6659 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 15:55:55.678255 6659 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 15:55:55.681248 6659 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:55.681331 6659 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:55.681366 6659 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 15:55:55.681374 6659 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 15:55:55.681388 6659 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 15:55:55.681426 6659 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 15:55:55.681421 6659 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:55.681519 6659 factory.go:656] Stopping watch factory\\\\nI0126 15:55:55.681560 6659 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:55.681572 6659 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 15:55:55.681581 6659 ovnkube.go:599] Stopped ovnkube\\\\nI0126 15:55:55.681663 6659 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 15:55:55.681759 6659 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.116734 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.136517 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.160581 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.177756 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.191795 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.191843 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.191851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.191866 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.191876 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.197879 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.213017 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:57 crc kubenswrapper[4880]: E0126 15:55:57.213193 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.213207 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:57 crc kubenswrapper[4880]: E0126 15:55:57.213563 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.214795 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:55:57Z is after 2025-08-24T17:21:41Z" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.295268 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.295303 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.295313 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.295329 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.295341 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.397873 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.397937 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.397948 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.397967 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.397980 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.441500 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 12:52:16.871940562 +0000 UTC Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.502167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.502232 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.502272 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.502296 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.502316 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.604662 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.604698 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.604707 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.604720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.604729 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.707413 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.707471 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.707482 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.707545 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.707573 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.811085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.811154 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.811167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.811190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.811202 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915743 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/2.log" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915785 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915805 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915813 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915827 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:57 crc kubenswrapper[4880]: I0126 15:55:57.915837 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:57Z","lastTransitionTime":"2026-01-26T15:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.019382 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.019458 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.019470 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.019488 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.019500 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.124286 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.124358 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.124373 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.124395 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.124409 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.213181 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.213248 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:55:58 crc kubenswrapper[4880]: E0126 15:55:58.213343 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:55:58 crc kubenswrapper[4880]: E0126 15:55:58.213522 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.227866 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.227936 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.227952 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.227972 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.227988 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.331046 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.331103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.331115 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.331222 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.331241 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.433720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.433776 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.433789 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.433807 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.433818 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.441880 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 20:21:43.92645002 +0000 UTC Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.536964 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.537000 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.537013 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.537030 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.537041 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.640344 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.640393 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.640411 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.640464 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.640485 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.744024 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.744072 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.744086 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.744103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.744117 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.847504 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.847554 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.847563 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.847579 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.847588 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.950899 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.950961 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.950974 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.950997 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:58 crc kubenswrapper[4880]: I0126 15:55:58.951020 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:58Z","lastTransitionTime":"2026-01-26T15:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.054722 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.055119 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.055275 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.055391 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.055559 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.159338 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.159404 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.159416 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.159465 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.159496 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.213350 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.213475 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:55:59 crc kubenswrapper[4880]: E0126 15:55:59.213926 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:55:59 crc kubenswrapper[4880]: E0126 15:55:59.214615 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.264041 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.264120 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.264136 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.264161 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.264178 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.368481 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.368541 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.368555 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.368576 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.368591 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.442509 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 22:22:29.66315262 +0000 UTC Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.472029 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.472076 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.472089 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.472106 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.472124 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.575458 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.575500 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.575518 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.575617 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.575634 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.678837 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.678894 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.678911 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.678932 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.678944 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.781671 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.781735 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.781750 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.781770 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.781782 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.884799 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.884844 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.884855 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.884874 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.884885 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.991596 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.991984 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.992004 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.992025 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:55:59 crc kubenswrapper[4880]: I0126 15:55:59.992039 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:55:59Z","lastTransitionTime":"2026-01-26T15:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.096087 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.096139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.096149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.096166 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.096179 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.199708 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.199769 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.199788 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.199805 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.199820 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.213608 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.213688 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:00 crc kubenswrapper[4880]: E0126 15:56:00.213803 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:00 crc kubenswrapper[4880]: E0126 15:56:00.213894 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.302653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.302783 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.302798 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.302825 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.302843 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.405747 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.405819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.405830 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.405848 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.405861 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.443256 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 03:21:18.30637478 +0000 UTC Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.508550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.508627 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.508642 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.508669 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.508683 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.612994 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.613053 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.613062 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.613107 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.613119 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.716190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.716253 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.716270 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.716290 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.716301 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.819241 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.819302 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.819312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.819333 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.819346 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.921799 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.921860 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.921874 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.921891 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:00 crc kubenswrapper[4880]: I0126 15:56:00.921901 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:00Z","lastTransitionTime":"2026-01-26T15:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.025498 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.025558 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.025575 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.025600 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.025615 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.127997 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.128050 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.128065 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.128086 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.128099 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.213712 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.213735 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:01 crc kubenswrapper[4880]: E0126 15:56:01.214082 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:01 crc kubenswrapper[4880]: E0126 15:56:01.214205 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.231697 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.231776 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.231812 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.231830 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.231842 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.335407 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.335476 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.335487 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.335505 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.335515 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.437514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.437564 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.437576 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.437590 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.437600 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.443717 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 01:24:54.503946763 +0000 UTC Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.542273 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.542339 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.542350 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.542373 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.542387 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.646081 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.646134 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.646144 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.646163 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.646174 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.749465 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.749556 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.749568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.749610 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.749625 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.852251 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.852336 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.852354 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.852375 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.852412 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.955916 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.955959 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.955969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.955990 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:01 crc kubenswrapper[4880]: I0126 15:56:01.956001 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:01Z","lastTransitionTime":"2026-01-26T15:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.059770 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.059822 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.059834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.059852 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.059864 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.164727 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.164819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.164832 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.164859 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.164874 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.213524 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.213524 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:02 crc kubenswrapper[4880]: E0126 15:56:02.213675 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:02 crc kubenswrapper[4880]: E0126 15:56:02.214730 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.268205 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.268277 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.268286 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.268303 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.268316 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.371462 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.371514 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.371524 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.371539 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.371549 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.444097 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 13:09:12.164927834 +0000 UTC Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.474785 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.474851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.474863 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.474883 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.474901 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.578063 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.578168 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.578181 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.578202 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.578215 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.681407 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.681546 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.681564 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.681584 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.681596 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.785780 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.785839 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.785850 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.785866 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.785876 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.889484 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.889536 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.889550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.889581 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.889637 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.993092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.993142 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.993152 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.993167 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:02 crc kubenswrapper[4880]: I0126 15:56:02.993178 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:02Z","lastTransitionTime":"2026-01-26T15:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.097724 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.097786 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.097801 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.097822 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.097835 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.201316 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.201396 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.201407 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.201424 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.201462 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.214053 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.214156 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:03 crc kubenswrapper[4880]: E0126 15:56:03.214224 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:03 crc kubenswrapper[4880]: E0126 15:56:03.214344 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.304892 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.304980 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.304997 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.305024 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.305039 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.408324 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.408395 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.408410 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.408453 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.408468 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.444668 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 11:44:53.858705255 +0000 UTC Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.511834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.511896 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.511910 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.511929 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.511943 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.615457 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.615512 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.615522 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.615537 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.615548 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.718795 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.718845 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.718858 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.718877 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.718895 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.822835 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.822906 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.822921 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.822943 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.822955 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.926597 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.926646 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.926657 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.926672 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.926683 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.976041 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.976085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.976097 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.976114 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.976124 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:03 crc kubenswrapper[4880]: E0126 15:56:03.993120 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:03Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.998482 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.998559 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.998574 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.998606 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:03 crc kubenswrapper[4880]: I0126 15:56:03.998622 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:03Z","lastTransitionTime":"2026-01-26T15:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.015418 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:04Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.022862 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.022963 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.022974 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.022993 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.023007 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.039832 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:04Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.044881 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.048635 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.048746 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.048775 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.048791 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.066930 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:04Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.072275 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.072333 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.072344 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.072393 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.072407 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.088956 4880 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T15:56:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0d360a7c-a5a7-4e03-b590-9896e616c72e\\\",\\\"systemUUID\\\":\\\"a6de2103-0bf5-4105-bc6f-45ed5bf76a5c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:04Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.089106 4880 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.091151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.091200 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.091210 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.091226 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.091238 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.194110 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.194157 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.194188 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.194207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.194218 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.213693 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.213839 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.213974 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.214194 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.297625 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.297689 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.297705 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.297723 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.297737 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.401353 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.401540 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.401551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.401568 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.401580 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.445043 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 15:02:24.484351918 +0000 UTC Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.504006 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.504050 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.504059 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.504076 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.504088 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.606366 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.606419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.606472 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.606504 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.606522 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.709545 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.709602 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.709617 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.709635 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.709650 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.813192 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.813234 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.813247 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.813263 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.813278 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.915528 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.915606 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.915616 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.915636 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.915648 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:04Z","lastTransitionTime":"2026-01-26T15:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:04 crc kubenswrapper[4880]: I0126 15:56:04.962124 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.962353 4880 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:56:04 crc kubenswrapper[4880]: E0126 15:56:04.962458 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs podName:216872c9-037e-486f-bf08-5e970a2bd5b1 nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.962416961 +0000 UTC m=+122.448145668 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs") pod "network-metrics-daemon-pvdts" (UID: "216872c9-037e-486f-bf08-5e970a2bd5b1") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.017848 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.017902 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.017917 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.017934 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.017947 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.120794 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.120855 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.120869 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.120887 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.120901 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.213733 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.213816 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:05 crc kubenswrapper[4880]: E0126 15:56:05.213907 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:05 crc kubenswrapper[4880]: E0126 15:56:05.214036 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.223702 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.223744 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.223754 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.223772 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.223782 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.326518 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.326577 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.326588 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.326605 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.326617 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.429720 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.429777 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.429788 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.429810 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.429822 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.445965 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:45:13.449757746 +0000 UTC Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.533126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.533190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.533199 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.533217 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.533229 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.637489 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.637565 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.637581 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.637604 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.637619 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.740743 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.740797 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.740809 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.740824 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.740835 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.843831 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.843896 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.843915 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.843941 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.843955 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.947430 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.947546 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.947566 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.947588 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:05 crc kubenswrapper[4880]: I0126 15:56:05.947602 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:05Z","lastTransitionTime":"2026-01-26T15:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.051221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.051270 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.051281 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.051299 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.051313 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.154421 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.154961 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.154980 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.155045 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.155062 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.214031 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.214115 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:06 crc kubenswrapper[4880]: E0126 15:56:06.214297 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:06 crc kubenswrapper[4880]: E0126 15:56:06.214532 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.236156 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258027 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258766 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258805 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258814 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258828 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.258838 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.278195 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.296079 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.320422 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.350920 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.362014 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.362078 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.362092 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.362112 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.362129 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.375912 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.394536 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.416043 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.430168 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.445425 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.446484 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 21:07:30.676765679 +0000 UTC Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.462850 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.464658 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.464697 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.464708 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.464723 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.464734 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.481070 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.498875 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.528669 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9164dafb83578a4e8cb0548a2f28e7241758400d806965d6dd8cfb24867517cf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:41Z\\\",\\\"message\\\":\\\"rk=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:5000, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0126 15:55:40.871524 6397 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI0126 15:55:40.871535 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871539 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0126 15:55:40.871552 6397 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-mpnc5\\\\nI0126 15:55:40.871549 6397 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-2p7jp\\\\nF0126 15:55:40.871561 6397 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:55Z\\\",\\\"message\\\":\\\"5:55:55.676520 6659 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 15:55:55.678224 6659 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 15:55:55.678255 6659 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 15:55:55.681248 6659 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:55.681331 6659 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:55.681366 6659 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 15:55:55.681374 6659 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 15:55:55.681388 6659 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 15:55:55.681426 6659 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 15:55:55.681421 6659 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:55.681519 6659 factory.go:656] Stopping watch factory\\\\nI0126 15:55:55.681560 6659 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:55.681572 6659 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 15:55:55.681581 6659 ovnkube.go:599] Stopped ovnkube\\\\nI0126 15:55:55.681663 6659 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 15:55:55.681759 6659 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.544279 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.561420 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:06Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.567863 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.567914 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.567923 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.567939 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.567951 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.671137 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.671197 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.671213 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.671232 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.671249 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.774551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.774605 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.774614 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.774632 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.774643 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.877712 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.877766 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.877780 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.877797 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.877809 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.980124 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.980179 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.980216 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.980233 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:06 crc kubenswrapper[4880]: I0126 15:56:06.980242 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:06Z","lastTransitionTime":"2026-01-26T15:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.084137 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.084223 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.084236 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.084252 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.084262 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.187833 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.187898 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.187911 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.187931 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.187947 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.213184 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.213187 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:07 crc kubenswrapper[4880]: E0126 15:56:07.213565 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:07 crc kubenswrapper[4880]: E0126 15:56:07.213871 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.243772 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.291345 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.291428 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.291458 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.291475 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.291488 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.395085 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.395128 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.395140 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.395154 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.395164 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.447640 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 06:01:53.374973476 +0000 UTC Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.497787 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.497834 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.497845 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.497861 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.497872 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.600736 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.600781 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.600793 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.600809 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.600818 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.703935 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.703993 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.704008 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.704027 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.704039 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.808054 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.808116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.808126 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.808144 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.808157 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.911307 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.911376 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.911390 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.911411 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:07 crc kubenswrapper[4880]: I0126 15:56:07.911424 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:07Z","lastTransitionTime":"2026-01-26T15:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.014748 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.014800 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.014822 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.014842 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.014855 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.118186 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.118234 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.118244 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.118259 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.118271 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.214138 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.214376 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:08 crc kubenswrapper[4880]: E0126 15:56:08.214666 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:08 crc kubenswrapper[4880]: E0126 15:56:08.214774 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.220548 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.220676 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.220691 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.220711 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.220724 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.324116 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.324207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.324221 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.324268 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.324283 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.429025 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.429089 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.429102 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.429120 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.429133 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.447954 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 15:43:49.723438946 +0000 UTC Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.532023 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.532089 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.532101 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.532121 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.532135 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.635190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.635251 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.635263 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.635280 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.635294 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.738235 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.738283 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.738295 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.738311 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.738322 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.841962 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.842020 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.842033 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.842050 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.842064 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.944766 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.944835 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.944848 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.944868 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:08 crc kubenswrapper[4880]: I0126 15:56:08.944879 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:08Z","lastTransitionTime":"2026-01-26T15:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.048922 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.048988 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.048999 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.049019 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.049042 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.110742 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.110956 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:57:13.110929001 +0000 UTC m=+158.596657718 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.111870 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.111946 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.112041 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112091 4880 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112228 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:57:13.112200923 +0000 UTC m=+158.597929650 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112231 4880 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112337 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 15:57:13.112291375 +0000 UTC m=+158.598053763 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112543 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112576 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112594 4880 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.112657 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 15:57:13.112643634 +0000 UTC m=+158.598372351 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.152008 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.152067 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.152081 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.152105 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.152118 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.213467 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.213571 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.213676 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.213961 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.214662 4880 scope.go:117] "RemoveContainer" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 15:56:09 crc kubenswrapper[4880]: E0126 15:56:09.214886 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2p7jp_openshift-ovn-kubernetes(c377ad66-a182-4cbf-9b29-253f3e547118)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.232890 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5697eee-6044-4d3d-823b-8300974b447f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9f760cbccedce41336851fc934005f63be48d0fbd138a81cd0b6458fbbaf2b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://149a2be2de817ba94c3f6cb5365bff399344238d167e29a3ca824d8fc9581018\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39f27d794705accea1c7c57b4b55cf05dd0077a375dad6fc2cbe3e44cf9e13a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.249096 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.255753 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.255818 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.255831 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.255847 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.255864 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.267070 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-h6f8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ddcd016a-fdad-4580-adaf-30451ac85f0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vz2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-h6f8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.287319 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbe43e2-b41e-4bbe-8bb3-c3389816a067\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d551fa38a07e147e4135dbb46ea26c8a06ac47804afb33430e788616a632976c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f6a336a066a79cf3920bf8972eecda3a4ee28f5ec65010206eb642d91681bd4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7af14c7ce33959eb69bc2350d6550d5b367bb252aa85b2c93d6a42f34d51aa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99336ddc867e2095eb92736279ea7fad47e8ab726d2d4928900b164086817c06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://502730be9fbc83f423b3ac863011d9e291fa714096ffd2f4a8fbe11d4103f33f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dd4a00666e30a814117341cc53dbe346b007de5703a2abe4763e0baf9ab9e2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://186dac83d1561fdf99095ec56adbb3fbb71d65515ad4e6ef1294cd666c01e150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kq76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-b2pvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.302041 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mpnc5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2af6c3-36e8-40a6-816b-52d5d64fe9ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbc310e2d56d8f5841f98dd50edc98494370f766ddba8f75015cf6e16736693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgqzb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mpnc5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.326720 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c377ad66-a182-4cbf-9b29-253f3e547118\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T15:55:55Z\\\",\\\"message\\\":\\\"5:55:55.676520 6659 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 15:55:55.678224 6659 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 15:55:55.678255 6659 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 15:55:55.681248 6659 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 15:55:55.681331 6659 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 15:55:55.681366 6659 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 15:55:55.681374 6659 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 15:55:55.681388 6659 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 15:55:55.681426 6659 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 15:55:55.681421 6659 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0126 15:55:55.681519 6659 factory.go:656] Stopping watch factory\\\\nI0126 15:55:55.681560 6659 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0126 15:55:55.681572 6659 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 15:55:55.681581 6659 ovnkube.go:599] Stopped ovnkube\\\\nI0126 15:55:55.681663 6659 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 15:55:55.681759 6659 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2p7jp_openshift-ovn-kubernetes(c377ad66-a182-4cbf-9b29-253f3e547118)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85xjf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2p7jp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.345149 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5beb13dc-5f14-4757-b9a0-01d074dfbd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b4d1a7ce7fd51231faded76a759ac9e53cb3ded2aa5c938d7dd4e808c421235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f229374fa75a5c2ebf256f08dedae14ccccfaa9685ec97553c7827b552091ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2z26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jrmsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359565 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-pvdts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216872c9-037e-486f-bf08-5e970a2bd5b1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgp6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-pvdts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359672 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359704 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359714 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359729 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.359741 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.376425 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8e17846-bd81-4e98-96e8-107f16f3a1fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90d7dd3edcf4866977855ab7f8521808da091c39150b4870863beab53ab01f2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://199638c0db57768114a763c71e9ce3c608b171f317dd391ea9d5a43b8f6daa48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbb90e484a41d3e1664925d8e19c94f13e2beadb50a673050b36687e16b1fcf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76dab33bdfeaa63fc4522157a5c5710b13a188459231d1724554a0f07e0a3786\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.399555 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f60d3693-6f46-4fd0-89bc-0d3ce9e2f935\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6d52682d8d63dbb7bbfb37e8ec3e8e33c5d202e2a9ff6b8589a15966ea683d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e105198778f920805554befbf8df70bd03c313f3cda928228595c40a91c3966a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9791e752e46f3b12715f8beebbe84367bd6014311b21ac47c13ffae707d1c56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15ca2ec49c0b701d86a2a86687031ca24529d4c98eca1c39871a3183ba42266a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edf0377b6bd8821563792c9d329e0ec2ba8d08e34cc96b1434fd452c657756aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0776888e982bbbc11351650436a5b372bf9fc5ff9687488b0176331b2394ead1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0776888e982bbbc11351650436a5b372bf9fc5ff9687488b0176331b2394ead1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad63ac306f03246543e6fe0b10d572f3324df9d0884206395e609dfb3b47280\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ad63ac306f03246543e6fe0b10d572f3324df9d0884206395e609dfb3b47280\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://714ae367da02b4746e89cbc307938f2231191e35e06d2cb252f94eb2094245fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://714ae367da02b4746e89cbc307938f2231191e35e06d2cb252f94eb2094245fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.416951 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0599f916-ef50-4fe0-ac05-8a41facb7a16\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:54:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T15:55:04Z\\\",\\\"message\\\":\\\"-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"kube-csr-signer_@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" (2025-02-24 05:54:36 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.809103019 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809135 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=6 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-controller-manager-operator_csr-signer-signer@1740376476\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:54:35 +0000 UTC to 2027-02-24 05:54:36 +0000 UTC (now=2026-01-26 15:55:04.80912061 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809153 1 tlsconfig.go:181] \\\\\\\"Loaded client CA\\\\\\\" index=7 certName=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"openshift-kube-apiserver-operator_aggregator-client-signer@1740376502\\\\\\\\\\\\\\\" [] issuer=\\\\\\\\\\\\\\\"\\\\u003cself\\\\u003e\\\\\\\\\\\\\\\" (2025-02-24 05:55:01 +0000 UTC to 2026-02-24 05:55:02 +0000 UTC (now=2026-01-26 15:55:04.80914065 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809276 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-982696073/tls.crt::/tmp/serving-cert-982696073/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1769442897\\\\\\\\\\\\\\\" (2026-01-26 15:54:56 +0000 UTC to 2026-02-25 15:54:57 +0000 UTC (now=2026-01-26 15:55:04.809264293 +0000 UTC))\\\\\\\"\\\\nI0126 15:55:04.809391 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1769442904\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1769442904\\\\\\\\\\\\\\\" (2026-01-26 14:55:04 +0000 UTC to 2027-01-26 14:55:04 +0000 UTC (now=2026-01-26 15:55:04.809378046 +0000 UTC))\\\\\\\"\\\\nF0126 15:55:04.918199 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:54:42Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T15:54:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T15:54:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:54:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.432314 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4a1ae0f858a676e9e76de05926c8e1cd99ae3d669e6282f903ccda84896b75c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.446089 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://febca1b971376e1e1ea0c25b03142e3e548b7da53e94a15ae8d1788cc8ebbd05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.448299 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:19:59.726565121 +0000 UTC Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462107 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462280 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462299 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462310 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.462280 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db202b43-2be9-4b6c-87b7-b2c5ef2df984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0635fcbb8ef9edc8de576962b842f224709cf14fd5a5110f64387dfef4ed999e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6d4mt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-glj8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.481218 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.498896 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://198d76ac5b45faf29eb321da32439030a4a2a690066522325ae370dd6016528a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c94c31f728a58eddeda07e3f45c9e3d5e7f9e631ec926c66fdfab2c915cdd35c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.515013 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.526663 4880 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ks4h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"512ee353-65e1-404a-9f3c-ee314b873f5e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T15:55:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7e3620716d57373c331d8b2d9e4d90a81351280eb265174e890682d200baf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T15:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rlwqg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T15:55:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ks4h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T15:56:09Z is after 2025-08-24T17:21:41Z" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.565404 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.565484 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.565495 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.565509 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.565518 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.668819 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.668886 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.668898 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.668918 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.668933 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.773792 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.773859 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.773869 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.773890 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.773901 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.876920 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.876977 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.876988 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.877006 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.877018 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.979181 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.979240 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.979257 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.979275 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:09 crc kubenswrapper[4880]: I0126 15:56:09.979295 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:09Z","lastTransitionTime":"2026-01-26T15:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.081751 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.081793 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.081803 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.081821 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.081834 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.184666 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.184755 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.184777 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.184797 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.184810 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.213464 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.213531 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.213676 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.213793 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.287508 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.287563 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.287583 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.287603 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.287616 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.325130 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.325468 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.325518 4880 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.325532 4880 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:56:10 crc kubenswrapper[4880]: E0126 15:56:10.325606 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 15:57:14.325587533 +0000 UTC m=+159.811316230 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.392614 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.393622 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.393713 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.393752 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.393771 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.448618 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 12:32:16.92978866 +0000 UTC Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.497782 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.497832 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.497842 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.497860 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.497876 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.602687 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.602817 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.602832 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.602852 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.602865 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.705778 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.705939 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.705958 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.706389 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.706702 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.811190 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.811262 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.811277 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.811312 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.811330 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.914426 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.914851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.914968 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.915103 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:10 crc kubenswrapper[4880]: I0126 15:56:10.915210 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:10Z","lastTransitionTime":"2026-01-26T15:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.019042 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.019111 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.019128 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.019149 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.019163 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.122151 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.122207 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.122222 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.122242 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.122260 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.213584 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.213606 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:11 crc kubenswrapper[4880]: E0126 15:56:11.213953 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:11 crc kubenswrapper[4880]: E0126 15:56:11.214075 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.225247 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.225295 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.225309 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.225329 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.225342 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.229917 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.327926 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.327970 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.327980 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.327995 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.328006 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.431299 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.431345 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.431354 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.431374 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.431384 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.449859 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 18:27:30.873518799 +0000 UTC Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.534676 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.534771 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.534786 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.534809 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.534824 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.637969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.638018 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.638026 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.638046 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.638057 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.741431 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.741517 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.741531 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.741551 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.741566 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.844378 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.844465 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.844480 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.844501 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.844515 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.947653 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.947704 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.947717 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.947735 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:11 crc kubenswrapper[4880]: I0126 15:56:11.947748 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:11Z","lastTransitionTime":"2026-01-26T15:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.050001 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.050050 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.050061 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.050077 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.050090 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.152489 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.152561 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.152570 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.152587 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.152596 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.213527 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.213529 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:12 crc kubenswrapper[4880]: E0126 15:56:12.213723 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:12 crc kubenswrapper[4880]: E0126 15:56:12.213890 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.255304 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.255355 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.255366 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.255383 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.255394 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.358743 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.358861 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.358879 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.358914 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.358929 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.450592 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 22:07:01.192330504 +0000 UTC Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.463550 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.463640 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.463654 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.463678 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.463715 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.566138 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.566202 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.566216 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.566239 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.566260 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.669728 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.669772 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.669786 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.669804 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.669819 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.772867 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.772953 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.772967 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.772985 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.772998 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.875782 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.875836 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.875851 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.875878 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.875892 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.978576 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.978650 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.978663 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.978692 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:12 crc kubenswrapper[4880]: I0126 15:56:12.978712 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:12Z","lastTransitionTime":"2026-01-26T15:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.082062 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.082129 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.082139 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.082156 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.082166 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.185417 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.185491 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.185505 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.185524 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.185536 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.213594 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.213579 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:13 crc kubenswrapper[4880]: E0126 15:56:13.213779 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:13 crc kubenswrapper[4880]: E0126 15:56:13.213874 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.288232 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.288299 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.288314 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.288332 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.288344 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.391642 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.391741 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.391764 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.391784 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.391796 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.451384 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 22:59:38.936042759 +0000 UTC Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.494419 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.494547 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.494561 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.494578 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.494593 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.606858 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.606956 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.606969 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.606991 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.607004 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.710539 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.710593 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.710606 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.710622 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.710635 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.814122 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.814174 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.814186 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.814203 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.814215 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.917108 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.917164 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.917180 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.917197 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:13 crc kubenswrapper[4880]: I0126 15:56:13.917210 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:13Z","lastTransitionTime":"2026-01-26T15:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.022231 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.022356 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.022377 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.022406 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.022455 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:14Z","lastTransitionTime":"2026-01-26T15:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.126206 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.126259 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.126289 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.126309 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.126319 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:14Z","lastTransitionTime":"2026-01-26T15:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.213340 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.213359 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:14 crc kubenswrapper[4880]: E0126 15:56:14.213717 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:14 crc kubenswrapper[4880]: E0126 15:56:14.213789 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.229204 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.229252 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.229263 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.229284 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.229299 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:14Z","lastTransitionTime":"2026-01-26T15:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.286036 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.286084 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.286098 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.286120 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.286132 4880 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T15:56:14Z","lastTransitionTime":"2026-01-26T15:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.350189 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j"] Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.351585 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.354711 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.356206 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.356231 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.356792 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.405510 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mpnc5" podStartSLOduration=65.405465195 podStartE2EDuration="1m5.405465195s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.375308376 +0000 UTC m=+99.861037093" watchObservedRunningTime="2026-01-26 15:56:14.405465195 +0000 UTC m=+99.891193902" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.424312 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jrmsz" podStartSLOduration=64.424256054 podStartE2EDuration="1m4.424256054s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.424209703 +0000 UTC m=+99.909938420" watchObservedRunningTime="2026-01-26 15:56:14.424256054 +0000 UTC m=+99.909984761" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.452188 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 16:02:39.492364665 +0000 UTC Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.452347 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.461244 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.461190839 podStartE2EDuration="42.461190839s" podCreationTimestamp="2026-01-26 15:55:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.460987483 +0000 UTC m=+99.946716200" watchObservedRunningTime="2026-01-26 15:56:14.461190839 +0000 UTC m=+99.946919556" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.462218 4880 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.478351 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.478411 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.478691 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.478775 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.478806 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.494279 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=7.494258287 podStartE2EDuration="7.494258287s" podCreationTimestamp="2026-01-26 15:56:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.493981211 +0000 UTC m=+99.979709928" watchObservedRunningTime="2026-01-26 15:56:14.494258287 +0000 UTC m=+99.979986994" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.516842 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.5168213 podStartE2EDuration="1m10.5168213s" podCreationTimestamp="2026-01-26 15:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.516787659 +0000 UTC m=+100.002516366" watchObservedRunningTime="2026-01-26 15:56:14.5168213 +0000 UTC m=+100.002550007" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.575177 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=3.575158078 podStartE2EDuration="3.575158078s" podCreationTimestamp="2026-01-26 15:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.574174503 +0000 UTC m=+100.059903210" watchObservedRunningTime="2026-01-26 15:56:14.575158078 +0000 UTC m=+100.060886785" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579589 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579648 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579676 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579707 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579734 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579747 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.579853 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.580925 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-service-ca\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.587845 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.602249 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5c5c09e-6260-4c4a-97ae-2d2a751736c9-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-d758j\" (UID: \"e5c5c09e-6260-4c4a-97ae-2d2a751736c9\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.659884 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2ks4h" podStartSLOduration=65.659857151 podStartE2EDuration="1m5.659857151s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.65942992 +0000 UTC m=+100.145158637" watchObservedRunningTime="2026-01-26 15:56:14.659857151 +0000 UTC m=+100.145585858" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.673707 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podStartSLOduration=65.673667509 podStartE2EDuration="1m5.673667509s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.672685905 +0000 UTC m=+100.158414612" watchObservedRunningTime="2026-01-26 15:56:14.673667509 +0000 UTC m=+100.159396216" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.675786 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.699709 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=66.699689156 podStartE2EDuration="1m6.699689156s" podCreationTimestamp="2026-01-26 15:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.698770733 +0000 UTC m=+100.184499450" watchObservedRunningTime="2026-01-26 15:56:14.699689156 +0000 UTC m=+100.185417863" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.758997 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-h6f8z" podStartSLOduration=65.758966327 podStartE2EDuration="1m5.758966327s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.737750547 +0000 UTC m=+100.223479244" watchObservedRunningTime="2026-01-26 15:56:14.758966327 +0000 UTC m=+100.244695044" Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.990196 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" event={"ID":"e5c5c09e-6260-4c4a-97ae-2d2a751736c9","Type":"ContainerStarted","Data":"eb1dab08fd85ab81ecaa45610048543deada595a05bbcf8954f71f8798ca323e"} Jan 26 15:56:14 crc kubenswrapper[4880]: I0126 15:56:14.990602 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" event={"ID":"e5c5c09e-6260-4c4a-97ae-2d2a751736c9","Type":"ContainerStarted","Data":"7f7bb6edfb3812df6d9b943a3fe65f19cc7c332632f98c3a74aa44c5a3ddf3ba"} Jan 26 15:56:15 crc kubenswrapper[4880]: I0126 15:56:15.006370 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-b2pvv" podStartSLOduration=66.006341111 podStartE2EDuration="1m6.006341111s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:14.758462055 +0000 UTC m=+100.244190762" watchObservedRunningTime="2026-01-26 15:56:15.006341111 +0000 UTC m=+100.492069818" Jan 26 15:56:15 crc kubenswrapper[4880]: I0126 15:56:15.006933 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-d758j" podStartSLOduration=66.006925655 podStartE2EDuration="1m6.006925655s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:15.005704625 +0000 UTC m=+100.491433342" watchObservedRunningTime="2026-01-26 15:56:15.006925655 +0000 UTC m=+100.492654362" Jan 26 15:56:15 crc kubenswrapper[4880]: I0126 15:56:15.213726 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:15 crc kubenswrapper[4880]: E0126 15:56:15.213926 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:15 crc kubenswrapper[4880]: I0126 15:56:15.213760 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:15 crc kubenswrapper[4880]: E0126 15:56:15.214104 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.000424 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/0.log" Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.000501 4880 generic.go:334] "Generic (PLEG): container finished" podID="ddcd016a-fdad-4580-adaf-30451ac85f0e" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" exitCode=1 Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.000541 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerDied","Data":"b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232"} Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.000994 4880 scope.go:117] "RemoveContainer" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.213135 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:16 crc kubenswrapper[4880]: I0126 15:56:16.213198 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:16 crc kubenswrapper[4880]: E0126 15:56:16.215362 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:16 crc kubenswrapper[4880]: E0126 15:56:16.215391 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:17 crc kubenswrapper[4880]: I0126 15:56:17.006323 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/0.log" Jan 26 15:56:17 crc kubenswrapper[4880]: I0126 15:56:17.006427 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerStarted","Data":"7963c3cc01b50b41810def1f17a183f3f03df1ec0d64b079bda8841b0e6f900c"} Jan 26 15:56:17 crc kubenswrapper[4880]: I0126 15:56:17.213119 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:17 crc kubenswrapper[4880]: I0126 15:56:17.213251 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:17 crc kubenswrapper[4880]: E0126 15:56:17.213471 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:17 crc kubenswrapper[4880]: E0126 15:56:17.213614 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:18 crc kubenswrapper[4880]: I0126 15:56:18.214091 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:18 crc kubenswrapper[4880]: I0126 15:56:18.214091 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:18 crc kubenswrapper[4880]: E0126 15:56:18.214388 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:18 crc kubenswrapper[4880]: E0126 15:56:18.214739 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:19 crc kubenswrapper[4880]: I0126 15:56:19.214117 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:19 crc kubenswrapper[4880]: I0126 15:56:19.214139 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:19 crc kubenswrapper[4880]: E0126 15:56:19.214587 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:19 crc kubenswrapper[4880]: E0126 15:56:19.214656 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:20 crc kubenswrapper[4880]: I0126 15:56:20.213713 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:20 crc kubenswrapper[4880]: E0126 15:56:20.214049 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:20 crc kubenswrapper[4880]: I0126 15:56:20.213711 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:20 crc kubenswrapper[4880]: E0126 15:56:20.214502 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:21 crc kubenswrapper[4880]: I0126 15:56:21.213552 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:21 crc kubenswrapper[4880]: E0126 15:56:21.213712 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:21 crc kubenswrapper[4880]: I0126 15:56:21.213566 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:21 crc kubenswrapper[4880]: E0126 15:56:21.213939 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:21 crc kubenswrapper[4880]: I0126 15:56:21.215054 4880 scope.go:117] "RemoveContainer" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 15:56:22 crc kubenswrapper[4880]: I0126 15:56:22.213350 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:22 crc kubenswrapper[4880]: I0126 15:56:22.213357 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:22 crc kubenswrapper[4880]: E0126 15:56:22.213557 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:22 crc kubenswrapper[4880]: E0126 15:56:22.213823 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.032360 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/2.log" Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.035396 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerStarted","Data":"662f18544e42fc1d98d1b06fe7e06c28caf59e4a88a8701f25a28195713bb3d9"} Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.036896 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.081073 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podStartSLOduration=74.081040853 podStartE2EDuration="1m14.081040853s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:23.077811805 +0000 UTC m=+108.563540512" watchObservedRunningTime="2026-01-26 15:56:23.081040853 +0000 UTC m=+108.566769560" Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.213837 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:23 crc kubenswrapper[4880]: I0126 15:56:23.213948 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:23 crc kubenswrapper[4880]: E0126 15:56:23.214078 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:23 crc kubenswrapper[4880]: E0126 15:56:23.214143 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:24 crc kubenswrapper[4880]: I0126 15:56:24.213703 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:24 crc kubenswrapper[4880]: I0126 15:56:24.213781 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:24 crc kubenswrapper[4880]: E0126 15:56:24.213893 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:24 crc kubenswrapper[4880]: E0126 15:56:24.214100 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:25 crc kubenswrapper[4880]: I0126 15:56:25.213843 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:25 crc kubenswrapper[4880]: I0126 15:56:25.213956 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:25 crc kubenswrapper[4880]: E0126 15:56:25.214006 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:25 crc kubenswrapper[4880]: E0126 15:56:25.214150 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:25 crc kubenswrapper[4880]: I0126 15:56:25.282049 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pvdts"] Jan 26 15:56:26 crc kubenswrapper[4880]: I0126 15:56:26.252080 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:26 crc kubenswrapper[4880]: E0126 15:56:26.252536 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:26 crc kubenswrapper[4880]: I0126 15:56:26.252633 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:26 crc kubenswrapper[4880]: E0126 15:56:26.252740 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:26 crc kubenswrapper[4880]: I0126 15:56:26.255297 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:26 crc kubenswrapper[4880]: E0126 15:56:26.255468 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:27 crc kubenswrapper[4880]: I0126 15:56:27.213940 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:27 crc kubenswrapper[4880]: E0126 15:56:27.214077 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.213606 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:28 crc kubenswrapper[4880]: E0126 15:56:28.213819 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-pvdts" podUID="216872c9-037e-486f-bf08-5e970a2bd5b1" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.213633 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.213613 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:28 crc kubenswrapper[4880]: E0126 15:56:28.213949 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 15:56:28 crc kubenswrapper[4880]: E0126 15:56:28.214114 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.800428 4880 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.800813 4880 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.857410 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q8tjj"] Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.858131 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.858421 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f9lnb"] Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.863891 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.863950 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.864304 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.864530 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.864613 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.869073 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8"] Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.869303 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.873618 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t"] Jan 26 15:56:28 crc kubenswrapper[4880]: I0126 15:56:28.874523 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.031579 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.032107 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.032721 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.036726 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.039118 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.039979 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.040400 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jtwk\" (UniqueName: \"kubernetes.io/projected/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-kube-api-access-6jtwk\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.052169 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.052656 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.052989 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.053172 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.053284 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.053428 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.053600 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054050 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.041203 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-client\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054232 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb8dd\" (UniqueName: \"kubernetes.io/projected/3d13c678-e761-49ad-8fa9-390e3be41292-kube-api-access-xb8dd\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054267 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054296 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d13c678-e761-49ad-8fa9-390e3be41292-audit-dir\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054316 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkcfq\" (UniqueName: \"kubernetes.io/projected/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-kube-api-access-rkcfq\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054338 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054362 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054412 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054427 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054477 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab46cd9-c24c-481b-99ed-c3449c91fbb6-serving-cert\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054509 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-config\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054532 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gp7n\" (UniqueName: \"kubernetes.io/projected/dab46cd9-c24c-481b-99ed-c3449c91fbb6-kube-api-access-6gp7n\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054553 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-service-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054571 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-images\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054610 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054627 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84cd2\" (UniqueName: \"kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054657 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054678 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054691 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054710 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-encryption-config\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054738 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-audit-policies\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054761 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-serving-cert\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.054781 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-config\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.055737 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.055791 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.055837 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.055895 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.055804 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.056042 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.056811 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.056969 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.056989 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.057130 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.057339 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.057610 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.057648 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.057800 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058105 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058601 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058758 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058867 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058909 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.058987 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.059024 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.063700 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.063964 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.071516 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.071958 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.072270 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.072672 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.072901 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.072898 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-2wj2q"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.073579 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6ft2z"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.074012 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.074654 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.074671 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.075093 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.075318 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.075925 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.077586 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.077878 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.078050 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.078321 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.078572 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.079184 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.080078 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.080954 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.081186 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.081368 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.081557 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.081751 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.090618 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg95b"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.091636 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.092136 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4vdpp"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.092763 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.093491 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.093814 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.108969 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h6x5b"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.110015 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.123197 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.125472 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qjrvl"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.126172 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wbm2m"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.126586 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.127552 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.127620 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.128705 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.135531 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.135960 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-d4nqf"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.136395 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.136807 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2296m"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.138982 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.139655 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.140912 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.141789 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.426415 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.435638 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-encryption-config\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436097 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436186 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-audit-policies\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436213 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-serving-cert\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436242 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-client\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436272 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436305 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-config\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436334 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436470 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jtwk\" (UniqueName: \"kubernetes.io/projected/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-kube-api-access-6jtwk\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436518 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-client\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436543 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb8dd\" (UniqueName: \"kubernetes.io/projected/3d13c678-e761-49ad-8fa9-390e3be41292-kube-api-access-xb8dd\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436576 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436611 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/506d74f6-5ffd-4569-98db-949e699a9f22-metrics-tls\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436651 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db5e6499-8551-4d89-a05e-936487c9aa56-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436689 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436715 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-service-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436871 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e873659e-c74c-45e8-ba2b-13a16071c117-machine-approver-tls\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436924 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.436955 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/140e65cc-7b24-4d1c-b558-07ca6f71a29c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.437196 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/506d74f6-5ffd-4569-98db-949e699a9f22-trusted-ca\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.437246 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d13c678-e761-49ad-8fa9-390e3be41292-audit-dir\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.439570 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442239 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkcfq\" (UniqueName: \"kubernetes.io/projected/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-kube-api-access-rkcfq\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442380 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db5e6499-8551-4d89-a05e-936487c9aa56-config\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442419 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442736 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442800 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443027 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55ae9b4d-9703-4852-8725-b8564ec87c82-metrics-tls\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443068 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-trusted-ca-bundle\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443096 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh8rw\" (UniqueName: \"kubernetes.io/projected/69b11523-e5c0-41b8-84d7-82cf9521626e-kube-api-access-dh8rw\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443273 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wwv\" (UniqueName: \"kubernetes.io/projected/c8272257-ba55-49eb-84c5-65cf2987c8ce-kube-api-access-j9wwv\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443304 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jngj6\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-kube-api-access-jngj6\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443458 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443498 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-trusted-ca\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443701 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-config\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443741 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443773 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443808 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443928 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xds7h\" (UniqueName: \"kubernetes.io/projected/e873659e-c74c-45e8-ba2b-13a16071c117-kube-api-access-xds7h\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443960 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4l5w\" (UniqueName: \"kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.443987 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444144 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b11523-e5c0-41b8-84d7-82cf9521626e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444241 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94edc90a-5195-4297-b015-feb66415fae5-serving-cert\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444274 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/140e65cc-7b24-4d1c-b558-07ca6f71a29c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444339 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444365 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444516 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444555 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444588 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444613 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444724 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444754 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/69b11523-e5c0-41b8-84d7-82cf9521626e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444803 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab46cd9-c24c-481b-99ed-c3449c91fbb6-serving-cert\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444847 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-config\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444901 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gp7n\" (UniqueName: \"kubernetes.io/projected/dab46cd9-c24c-481b-99ed-c3449c91fbb6-kube-api-access-6gp7n\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444929 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-service-ca\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.444989 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-oauth-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.445023 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.445071 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db5e6499-8551-4d89-a05e-936487c9aa56-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.445103 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.446648 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.450385 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q8tjj"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.450647 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.452486 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.445203 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-service-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.452817 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-images\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.453061 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-auth-proxy-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.442280 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.476360 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.476509 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.476759 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.477164 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.477555 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.477629 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-config\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.478067 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-images\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.478543 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.478601 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.478634 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.478788 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479236 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479277 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479322 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479487 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479565 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d13c678-e761-49ad-8fa9-390e3be41292-audit-dir\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479620 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479719 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-audit-policies\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479746 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.453200 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-serving-cert\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479770 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479803 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxrbd\" (UniqueName: \"kubernetes.io/projected/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-kube-api-access-qxrbd\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479833 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnmm9\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-kube-api-access-cnmm9\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479859 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479871 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479915 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjt4\" (UniqueName: \"kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479942 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479961 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmfk5\" (UniqueName: \"kubernetes.io/projected/55ae9b4d-9703-4852-8725-b8564ec87c82-kube-api-access-cmfk5\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479983 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.480007 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.481672 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.481897 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.479999 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.482685 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ssxn4"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483192 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84cd2\" (UniqueName: \"kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483368 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-config\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483540 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483682 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483805 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483934 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484049 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484186 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-oauth-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484297 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484411 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t77qh\" (UniqueName: \"kubernetes.io/projected/94edc90a-5195-4297-b015-feb66415fae5-kube-api-access-t77qh\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484579 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140e65cc-7b24-4d1c-b558-07ca6f71a29c-config\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484696 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9xrg\" (UniqueName: \"kubernetes.io/projected/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-kube-api-access-x9xrg\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484833 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484952 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.491405 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.492720 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.483763 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.482161 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.484076 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.493741 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.494399 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.495764 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.482778 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.495770 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-service-ca-bundle\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.496573 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.496612 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.496948 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.496844 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.500593 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab46cd9-c24c-481b-99ed-c3449c91fbb6-config\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.497011 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.497933 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.498000 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.498103 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.498151 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.503783 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.505649 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.505702 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.505857 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.506033 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.506092 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.506138 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.513900 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.513948 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.513984 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.514074 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.516692 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.522510 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.525179 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.496996 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.525283 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab46cd9-c24c-481b-99ed-c3449c91fbb6-serving-cert\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.526381 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.527327 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.527597 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.527882 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.530083 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.530427 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.531780 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-encryption-config\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.532380 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-etcd-client\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.533724 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534060 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534242 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534456 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534615 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534739 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534855 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.534977 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535059 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d13c678-e761-49ad-8fa9-390e3be41292-serving-cert\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535208 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535420 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f9lnb"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535483 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535583 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535738 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.535868 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.536233 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.536279 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.536304 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.640197 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.643571 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:29 crc kubenswrapper[4880]: I0126 15:56:29.648736 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.653546 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.653713 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.653564 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654059 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654238 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654328 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654381 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-service-ca\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654425 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-oauth-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654489 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654537 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654558 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654572 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db5e6499-8551-4d89-a05e-936487c9aa56-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654604 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-auth-proxy-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654636 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-serving-cert\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654669 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxrbd\" (UniqueName: \"kubernetes.io/projected/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-kube-api-access-qxrbd\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654696 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnmm9\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-kube-api-access-cnmm9\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654722 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654753 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654772 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjt4\" (UniqueName: \"kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654828 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmfk5\" (UniqueName: \"kubernetes.io/projected/55ae9b4d-9703-4852-8725-b8564ec87c82-kube-api-access-cmfk5\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654860 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654886 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654933 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-config\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654958 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654989 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.654991 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655163 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655216 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655263 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-oauth-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655298 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655342 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t77qh\" (UniqueName: \"kubernetes.io/projected/94edc90a-5195-4297-b015-feb66415fae5-kube-api-access-t77qh\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655378 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140e65cc-7b24-4d1c-b558-07ca6f71a29c-config\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655410 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9xrg\" (UniqueName: \"kubernetes.io/projected/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-kube-api-access-x9xrg\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655475 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655531 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-client\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655556 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655588 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655648 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655658 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-service-ca\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655683 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/506d74f6-5ffd-4569-98db-949e699a9f22-metrics-tls\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655714 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db5e6499-8551-4d89-a05e-936487c9aa56-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655745 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655779 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-service-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655815 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e873659e-c74c-45e8-ba2b-13a16071c117-machine-approver-tls\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655849 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/140e65cc-7b24-4d1c-b558-07ca6f71a29c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655916 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-oauth-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.655878 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/506d74f6-5ffd-4569-98db-949e699a9f22-trusted-ca\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656018 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db5e6499-8551-4d89-a05e-936487c9aa56-config\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656049 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656083 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656137 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55ae9b4d-9703-4852-8725-b8564ec87c82-metrics-tls\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656295 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-trusted-ca-bundle\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656326 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh8rw\" (UniqueName: \"kubernetes.io/projected/69b11523-e5c0-41b8-84d7-82cf9521626e-kube-api-access-dh8rw\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656367 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wwv\" (UniqueName: \"kubernetes.io/projected/c8272257-ba55-49eb-84c5-65cf2987c8ce-kube-api-access-j9wwv\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656398 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jngj6\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-kube-api-access-jngj6\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656620 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-trusted-ca\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656659 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-config\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656687 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656717 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656859 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656891 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xds7h\" (UniqueName: \"kubernetes.io/projected/e873659e-c74c-45e8-ba2b-13a16071c117-kube-api-access-xds7h\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656918 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4l5w\" (UniqueName: \"kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.656950 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657056 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b11523-e5c0-41b8-84d7-82cf9521626e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657111 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94edc90a-5195-4297-b015-feb66415fae5-serving-cert\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657156 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/140e65cc-7b24-4d1c-b558-07ca6f71a29c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657196 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657312 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657353 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657383 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.657612 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.658916 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/69b11523-e5c0-41b8-84d7-82cf9521626e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.659756 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/69b11523-e5c0-41b8-84d7-82cf9521626e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.659978 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.666277 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.667768 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.667906 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.667988 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.668167 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.668312 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:29.671106 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.020108 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.027833 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-service-ca\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.036970 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.039215 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.040257 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.040856 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.040934 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.041301 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.041731 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w25jq"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.042059 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.042549 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.042969 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.043781 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.045429 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.046018 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.046265 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.047215 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.047430 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vmcgp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.048118 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.048534 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.049379 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.050350 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.051777 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.052418 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.052593 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.053088 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.054808 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg95b"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.055905 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/506d74f6-5ffd-4569-98db-949e699a9f22-trusted-ca\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.056566 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2wj2q"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.056764 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.057092 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.057325 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/506d74f6-5ffd-4569-98db-949e699a9f22-metrics-tls\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.057615 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.057725 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.058063 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-etcd-client\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.058262 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-oauth-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.058816 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e873659e-c74c-45e8-ba2b-13a16071c117-auth-proxy-config\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.059052 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6ft2z"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.060895 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.062545 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.062652 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4vdpp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.062656 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.062775 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.062786 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140e65cc-7b24-4d1c-b558-07ca6f71a29c-config\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.063737 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.064159 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-serving-cert\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.064476 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.064847 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.065629 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.065992 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.066649 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.067276 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.067494 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.066657 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-trusted-ca\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.067957 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-config\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.067877 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.068728 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-trusted-ca-bundle\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.069427 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94edc90a-5195-4297-b015-feb66415fae5-config\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.070204 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.070516 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.070707 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c8272257-ba55-49eb-84c5-65cf2987c8ce-console-config\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.080480 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.080850 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081065 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081257 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081508 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081670 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081691 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081745 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.081971 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.082137 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.082278 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.082314 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.082315 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.084469 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.084627 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.084828 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.085273 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db5e6499-8551-4d89-a05e-936487c9aa56-config\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.085680 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-serving-cert\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.086043 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.087704 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.089634 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.089966 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.090750 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.092500 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.092691 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.093055 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.094004 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kwks8"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.095697 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.095733 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.099174 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.101957 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.104775 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.110820 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84cd2\" (UniqueName: \"kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2\") pod \"route-controller-manager-6576b87f9c-lmp9b\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.112726 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db5e6499-8551-4d89-a05e-936487c9aa56-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.113574 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e873659e-c74c-45e8-ba2b-13a16071c117-machine-approver-tls\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.127368 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94edc90a-5195-4297-b015-feb66415fae5-serving-cert\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.150681 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b11523-e5c0-41b8-84d7-82cf9521626e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.151249 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.151767 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkcfq\" (UniqueName: \"kubernetes.io/projected/d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1-kube-api-access-rkcfq\") pod \"openshift-apiserver-operator-796bbdcf4f-76rr8\" (UID: \"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.152528 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.153194 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.154338 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.155036 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.155334 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.156157 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/140e65cc-7b24-4d1c-b558-07ca6f71a29c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.156164 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gp7n\" (UniqueName: \"kubernetes.io/projected/dab46cd9-c24c-481b-99ed-c3449c91fbb6-kube-api-access-6gp7n\") pod \"authentication-operator-69f744f599-q8tjj\" (UID: \"dab46cd9-c24c-481b-99ed-c3449c91fbb6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.157181 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jtwk\" (UniqueName: \"kubernetes.io/projected/e96b5221-3d1f-4f21-bb0f-4b01373ffdb0-kube-api-access-6jtwk\") pod \"machine-api-operator-5694c8668f-f9lnb\" (UID: \"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.158074 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/55ae9b4d-9703-4852-8725-b8564ec87c82-metrics-tls\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.158508 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.161623 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.162086 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.165246 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.166585 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qjrvl"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.181675 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.189483 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb8dd\" (UniqueName: \"kubernetes.io/projected/3d13c678-e761-49ad-8fa9-390e3be41292-kube-api-access-xb8dd\") pod \"apiserver-7bbb656c7d-szk2t\" (UID: \"3d13c678-e761-49ad-8fa9-390e3be41292\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.200612 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.170262 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.322973 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.332855 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.332991 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.334915 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.334995 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.336781 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.336797 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.338530 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.340846 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.343323 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.347672 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.347728 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.347789 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.347818 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wbm2m"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.349421 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h6x5b"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.355412 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ssxn4"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.363599 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.365122 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.365257 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.369009 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.370752 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.372552 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.374900 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.376061 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.377049 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.380643 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.382674 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kwks8"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.383716 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.383962 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w25jq"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.384783 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-gmdcv"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.386058 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.386150 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.390020 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pxxnr"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.391745 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gmdcv"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.391891 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.394603 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pxxnr"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.395472 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vmcgp"] Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.405207 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.424881 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.441681 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.461866 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.482154 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.483843 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.777145 4880 request.go:700] Waited for 1.122742221s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.780687 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.788172 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.788494 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.788701 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.788991 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.789120 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.790872 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.804804 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/140e65cc-7b24-4d1c-b558-07ca6f71a29c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plrb8\" (UID: \"140e65cc-7b24-4d1c-b558-07ca6f71a29c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.809766 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.812951 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9xrg\" (UniqueName: \"kubernetes.io/projected/62b30286-ebfb-4bd1-a781-dfd2f6ab107e-kube-api-access-x9xrg\") pod \"cluster-samples-operator-665b6dd947-8kg54\" (UID: \"62b30286-ebfb-4bd1-a781-dfd2f6ab107e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.815819 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.817605 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxrbd\" (UniqueName: \"kubernetes.io/projected/16aea370-960e-4bc7-a90c-fe7c7e9e6a4a-kube-api-access-qxrbd\") pod \"etcd-operator-b45778765-qjrvl\" (UID: \"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.818146 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh8rw\" (UniqueName: \"kubernetes.io/projected/69b11523-e5c0-41b8-84d7-82cf9521626e-kube-api-access-dh8rw\") pod \"openshift-config-operator-7777fb866f-cg95b\" (UID: \"69b11523-e5c0-41b8-84d7-82cf9521626e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:30 crc kubenswrapper[4880]: I0126 15:56:30.818770 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4l5w\" (UniqueName: \"kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w\") pod \"controller-manager-879f6c89f-4xvtl\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.013793 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.015261 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.018957 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.019001 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjt4\" (UniqueName: \"kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4\") pod \"oauth-openshift-558db77b4-pxfwl\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.020076 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.020425 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.021018 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-serving-cert\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024554 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/668172a7-5728-423d-a32e-a0737d06dbe9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024598 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-node-pullsecrets\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024619 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cd4170db-b57f-4f72-a59b-242d9227f742-signing-key\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024640 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/668172a7-5728-423d-a32e-a0737d06dbe9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024702 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8v7z\" (UniqueName: \"kubernetes.io/projected/d20ddb1e-869d-4625-a781-c2b35369fca5-kube-api-access-z8v7z\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024769 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-cert\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024785 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-webhook-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024819 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-service-ca-bundle\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024839 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024888 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024975 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j4q\" (UniqueName: \"kubernetes.io/projected/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-kube-api-access-f9j4q\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025007 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025037 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920d5f22-6fa6-4af8-8a47-554a8cffab54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025065 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4prz\" (UniqueName: \"kubernetes.io/projected/c322b39b-fdab-4043-bfa5-3408acab0d10-kube-api-access-r4prz\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025106 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-srv-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025122 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdb9f\" (UniqueName: \"kubernetes.io/projected/f972ba6d-640d-462e-9dc0-f8ee5454691e-kube-api-access-fdb9f\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025138 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdw82\" (UniqueName: \"kubernetes.io/projected/c996fc87-b129-4efb-9355-f355726ea4a7-kube-api-access-pdw82\") pod \"migrator-59844c95c7-km4zv\" (UID: \"c996fc87-b129-4efb-9355-f355726ea4a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025215 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025238 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025263 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025332 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025383 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/24d4b53c-609a-419f-803d-78e4b13b9afd-tmpfs\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025414 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-config-volume\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025437 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-metrics-tls\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025492 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025580 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8kgz\" (UniqueName: \"kubernetes.io/projected/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-kube-api-access-c8kgz\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025820 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.025908 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920d5f22-6fa6-4af8-8a47-554a8cffab54-proxy-tls\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.026006 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.020744 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024314 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.020849 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.023883 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.023988 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024058 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024119 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.024148 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.027297 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-client\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.027347 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgwmb\" (UniqueName: \"kubernetes.io/projected/24d4b53c-609a-419f-803d-78e4b13b9afd-kube-api-access-wgwmb\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.027503 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-audit\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.027550 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: E0126 15:56:31.027683 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:31.52766273 +0000 UTC m=+117.013391437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.028211 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-srv-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.028951 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cd4170db-b57f-4f72-a59b-242d9227f742-signing-cabundle\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.028988 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.029522 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c322b39b-fdab-4043-bfa5-3408acab0d10-config\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.029872 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-default-certificate\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.029909 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnmb\" (UniqueName: \"kubernetes.io/projected/cd4170db-b57f-4f72-a59b-242d9227f742-kube-api-access-jjnmb\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.029927 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wcz6\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.030625 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7k6x\" (UniqueName: \"kubernetes.io/projected/c0fb73bb-2bd2-460f-9d46-a1b44680e187-kube-api-access-b7k6x\") pod \"downloads-7954f5f757-2wj2q\" (UID: \"c0fb73bb-2bd2-460f-9d46-a1b44680e187\") " pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.031305 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-stats-auth\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.031402 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.031452 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd864\" (UniqueName: \"kubernetes.io/projected/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-kube-api-access-gd864\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.033847 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn9ww\" (UniqueName: \"kubernetes.io/projected/31b5147c-2b03-40d2-bb73-4c0752279b1c-kube-api-access-hn9ww\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.034078 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/668172a7-5728-423d-a32e-a0737d06dbe9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.034786 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-metrics-certs\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.034895 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5fbc\" (UniqueName: \"kubernetes.io/projected/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-kube-api-access-z5fbc\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.034937 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035201 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035241 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dslb\" (UniqueName: \"kubernetes.io/projected/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-kube-api-access-8dslb\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035299 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035357 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-encryption-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035549 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkn6s\" (UniqueName: \"kubernetes.io/projected/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-kube-api-access-qkn6s\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035618 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-image-import-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.035662 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c322b39b-fdab-4043-bfa5-3408acab0d10-serving-cert\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036021 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036094 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-apiservice-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036199 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036748 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59ckh\" (UniqueName: \"kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036876 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-audit-dir\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.036992 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtjvr\" (UniqueName: \"kubernetes.io/projected/920d5f22-6fa6-4af8-8a47-554a8cffab54-kube-api-access-xtjvr\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.063230 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmfk5\" (UniqueName: \"kubernetes.io/projected/55ae9b4d-9703-4852-8725-b8564ec87c82-kube-api-access-cmfk5\") pod \"dns-operator-744455d44c-wbm2m\" (UID: \"55ae9b4d-9703-4852-8725-b8564ec87c82\") " pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.063650 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.278080 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.280663 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.281657 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282044 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkn6s\" (UniqueName: \"kubernetes.io/projected/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-kube-api-access-qkn6s\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282112 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-image-import-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282170 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-csi-data-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282197 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c322b39b-fdab-4043-bfa5-3408acab0d10-serving-cert\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282221 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282297 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-certs\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282348 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282426 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-apiservice-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282499 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-registration-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282556 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282587 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-plugins-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282637 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59ckh\" (UniqueName: \"kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282706 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-audit-dir\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282767 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3638c1-682f-4bac-a083-ea8ecb9763a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282847 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtjvr\" (UniqueName: \"kubernetes.io/projected/920d5f22-6fa6-4af8-8a47-554a8cffab54-kube-api-access-xtjvr\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282890 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-node-bootstrap-token\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282933 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-serving-cert\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282960 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/668172a7-5728-423d-a32e-a0737d06dbe9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.282991 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-node-pullsecrets\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283011 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cd4170db-b57f-4f72-a59b-242d9227f742-signing-key\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283038 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/668172a7-5728-423d-a32e-a0737d06dbe9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283101 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8v7z\" (UniqueName: \"kubernetes.io/projected/d20ddb1e-869d-4625-a781-c2b35369fca5-kube-api-access-z8v7z\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283191 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283270 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-cert\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283309 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-webhook-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283332 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-service-ca-bundle\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283362 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283387 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-socket-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283469 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-677lc\" (UniqueName: \"kubernetes.io/projected/c631c684-80bc-4f57-ae38-bb96c0db15d1-kube-api-access-677lc\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283557 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283615 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j4q\" (UniqueName: \"kubernetes.io/projected/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-kube-api-access-f9j4q\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283642 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283690 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a651d0-6847-4104-9478-f46f6d8d7c7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283742 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920d5f22-6fa6-4af8-8a47-554a8cffab54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283776 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4prz\" (UniqueName: \"kubernetes.io/projected/c322b39b-fdab-4043-bfa5-3408acab0d10-kube-api-access-r4prz\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283810 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zxk\" (UniqueName: \"kubernetes.io/projected/a91a0129-b4bd-4acd-bceb-3091d5714089-kube-api-access-c6zxk\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283851 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-srv-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283894 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283924 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdb9f\" (UniqueName: \"kubernetes.io/projected/f972ba6d-640d-462e-9dc0-f8ee5454691e-kube-api-access-fdb9f\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283950 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdw82\" (UniqueName: \"kubernetes.io/projected/c996fc87-b129-4efb-9355-f355726ea4a7-kube-api-access-pdw82\") pod \"migrator-59844c95c7-km4zv\" (UID: \"c996fc87-b129-4efb-9355-f355726ea4a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.283980 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284007 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284024 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284041 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzgv7\" (UniqueName: \"kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284063 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/24d4b53c-609a-419f-803d-78e4b13b9afd-tmpfs\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284087 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-config-volume\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284121 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-metrics-tls\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284147 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffm7h\" (UniqueName: \"kubernetes.io/projected/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-kube-api-access-ffm7h\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284187 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284230 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8kgz\" (UniqueName: \"kubernetes.io/projected/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-kube-api-access-c8kgz\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284303 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284352 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920d5f22-6fa6-4af8-8a47-554a8cffab54-proxy-tls\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284391 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff3638c1-682f-4bac-a083-ea8ecb9763a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284511 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-client\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284605 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-audit\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284651 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284681 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgwmb\" (UniqueName: \"kubernetes.io/projected/24d4b53c-609a-419f-803d-78e4b13b9afd-kube-api-access-wgwmb\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284706 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-srv-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284751 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-images\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284777 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cd4170db-b57f-4f72-a59b-242d9227f742-signing-cabundle\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284799 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284826 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kflfw\" (UniqueName: \"kubernetes.io/projected/76a651d0-6847-4104-9478-f46f6d8d7c7a-kube-api-access-kflfw\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284883 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-default-certificate\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284915 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnmb\" (UniqueName: \"kubernetes.io/projected/cd4170db-b57f-4f72-a59b-242d9227f742-kube-api-access-jjnmb\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284940 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c322b39b-fdab-4043-bfa5-3408acab0d10-config\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.284979 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wcz6\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285004 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w6dz\" (UniqueName: \"kubernetes.io/projected/ff3638c1-682f-4bac-a083-ea8ecb9763a4-kube-api-access-5w6dz\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285029 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7k6x\" (UniqueName: \"kubernetes.io/projected/c0fb73bb-2bd2-460f-9d46-a1b44680e187-kube-api-access-b7k6x\") pod \"downloads-7954f5f757-2wj2q\" (UID: \"c0fb73bb-2bd2-460f-9d46-a1b44680e187\") " pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285044 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-mountpoint-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285062 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-stats-auth\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285081 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285097 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd864\" (UniqueName: \"kubernetes.io/projected/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-kube-api-access-gd864\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285114 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn9ww\" (UniqueName: \"kubernetes.io/projected/31b5147c-2b03-40d2-bb73-4c0752279b1c-kube-api-access-hn9ww\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285132 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/668172a7-5728-423d-a32e-a0737d06dbe9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285163 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-metrics-certs\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285185 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285211 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5fbc\" (UniqueName: \"kubernetes.io/projected/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-kube-api-access-z5fbc\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285248 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285305 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285327 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dslb\" (UniqueName: \"kubernetes.io/projected/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-kube-api-access-8dslb\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285350 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285376 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a91a0129-b4bd-4acd-bceb-3091d5714089-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.285412 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-encryption-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.287373 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.289424 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/24d4b53c-609a-419f-803d-78e4b13b9afd-tmpfs\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: E0126 15:56:31.290246 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:31.790218135 +0000 UTC m=+117.275946842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.290426 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-encryption-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.310414 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-image-import-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.310546 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.310568 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-apiservice-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.314672 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.317997 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.319421 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-node-pullsecrets\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.319675 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d20ddb1e-869d-4625-a781-c2b35369fca5-audit-dir\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.342899 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-serving-ca\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.343761 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-etcd-client\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.344305 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d20ddb1e-869d-4625-a781-c2b35369fca5-serving-cert\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.346226 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.349292 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-stats-auth\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.343028 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/668172a7-5728-423d-a32e-a0737d06dbe9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.349563 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c322b39b-fdab-4043-bfa5-3408acab0d10-serving-cert\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.352315 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920d5f22-6fa6-4af8-8a47-554a8cffab54-proxy-tls\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.355668 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wwv\" (UniqueName: \"kubernetes.io/projected/c8272257-ba55-49eb-84c5-65cf2987c8ce-kube-api-access-j9wwv\") pod \"console-f9d7485db-6ft2z\" (UID: \"c8272257-ba55-49eb-84c5-65cf2987c8ce\") " pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.356408 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.365600 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.368827 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-srv-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.370077 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.370830 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-service-ca-bundle\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.373766 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/24d4b53c-609a-419f-803d-78e4b13b9afd-webhook-cert\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.381731 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.382176 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920d5f22-6fa6-4af8-8a47-554a8cffab54-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.382613 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/668172a7-5728-423d-a32e-a0737d06dbe9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.383828 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c322b39b-fdab-4043-bfa5-3408acab0d10-config\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.385613 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.388896 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-config\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.390571 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-default-certificate\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.391153 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.391771 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-csi-data-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.391928 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-csi-data-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.391950 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d20ddb1e-869d-4625-a781-c2b35369fca5-audit\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.391988 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.392310 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-registration-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.393021 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-registration-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.393072 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-plugins-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.393160 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-plugins-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.393509 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3638c1-682f-4bac-a083-ea8ecb9763a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.394330 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff3638c1-682f-4bac-a083-ea8ecb9763a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.394459 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-node-bootstrap-token\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395200 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395428 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a651d0-6847-4104-9478-f46f6d8d7c7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395523 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zxk\" (UniqueName: \"kubernetes.io/projected/a91a0129-b4bd-4acd-bceb-3091d5714089-kube-api-access-c6zxk\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395604 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzgv7\" (UniqueName: \"kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395666 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffm7h\" (UniqueName: \"kubernetes.io/projected/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-kube-api-access-ffm7h\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395773 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff3638c1-682f-4bac-a083-ea8ecb9763a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395846 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395989 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.396086 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.396363 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-metrics-certs\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.396420 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.395998 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-images\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.396541 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kflfw\" (UniqueName: \"kubernetes.io/projected/76a651d0-6847-4104-9478-f46f6d8d7c7a-kube-api-access-kflfw\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.396976 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-images\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: E0126 15:56:31.397431 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:31.897406889 +0000 UTC m=+117.383135596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.397926 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/76a651d0-6847-4104-9478-f46f6d8d7c7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.398955 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.399650 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.400013 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.403526 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cd4170db-b57f-4f72-a59b-242d9227f742-signing-key\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.404062 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db5e6499-8551-4d89-a05e-936487c9aa56-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mkbhn\" (UID: \"db5e6499-8551-4d89-a05e-936487c9aa56\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.406858 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.406928 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.407045 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.407527 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.408645 4880 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.415869 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cd4170db-b57f-4f72-a59b-242d9227f742-signing-cabundle\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.486366 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f972ba6d-640d-462e-9dc0-f8ee5454691e-profile-collector-cert\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.486776 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.511490 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.511517 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.511783 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.513062 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.514975 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.690654 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-config-volume\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.711192 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.712766 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8kgz\" (UniqueName: \"kubernetes.io/projected/9340b971-8db7-4e70-bbc9-0e4d7292b4e1-kube-api-access-c8kgz\") pod \"router-default-5444994796-d4nqf\" (UID: \"9340b971-8db7-4e70-bbc9-0e4d7292b4e1\") " pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.712754 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31b5147c-2b03-40d2-bb73-4c0752279b1c-srv-cert\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.723938 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-certs\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: E0126 15:56:31.724173 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:32.224129256 +0000 UTC m=+117.709857973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.725113 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.726148 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.726932 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t77qh\" (UniqueName: \"kubernetes.io/projected/94edc90a-5195-4297-b015-feb66415fae5-kube-api-access-t77qh\") pod \"console-operator-58897d9998-4vdpp\" (UID: \"94edc90a-5195-4297-b015-feb66415fae5\") " pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.729067 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-socket-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.729496 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-677lc\" (UniqueName: \"kubernetes.io/projected/c631c684-80bc-4f57-ae38-bb96c0db15d1-kube-api-access-677lc\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.729896 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.729315 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-socket-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730009 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-node-bootstrap-token\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730220 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w6dz\" (UniqueName: \"kubernetes.io/projected/ff3638c1-682f-4bac-a083-ea8ecb9763a4-kube-api-access-5w6dz\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730282 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-mountpoint-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730405 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730511 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a91a0129-b4bd-4acd-bceb-3091d5714089-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:31 crc kubenswrapper[4880]: E0126 15:56:31.730632 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:32.230609805 +0000 UTC m=+117.716338562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.730762 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c631c684-80bc-4f57-ae38-bb96c0db15d1-mountpoint-dir\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.732213 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.732326 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/76a651d0-6847-4104-9478-f46f6d8d7c7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.750303 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff3638c1-682f-4bac-a083-ea8ecb9763a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.750330 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jngj6\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-kube-api-access-jngj6\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.750398 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1296f6c-3717-4a9c-abc1-d27c4b98dd10-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jtztn\" (UID: \"e1296f6c-3717-4a9c-abc1-d27c4b98dd10\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.753077 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xds7h\" (UniqueName: \"kubernetes.io/projected/e873659e-c74c-45e8-ba2b-13a16071c117-kube-api-access-xds7h\") pod \"machine-approver-56656f9798-46rpj\" (UID: \"e873659e-c74c-45e8-ba2b-13a16071c117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.753375 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j4q\" (UniqueName: \"kubernetes.io/projected/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-kube-api-access-f9j4q\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.753612 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtjvr\" (UniqueName: \"kubernetes.io/projected/920d5f22-6fa6-4af8-8a47-554a8cffab54-kube-api-access-xtjvr\") pod \"machine-config-controller-84d6567774-j9b2x\" (UID: \"920d5f22-6fa6-4af8-8a47-554a8cffab54\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.753658 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn9ww\" (UniqueName: \"kubernetes.io/projected/31b5147c-2b03-40d2-bb73-4c0752279b1c-kube-api-access-hn9ww\") pod \"olm-operator-6b444d44fb-52d79\" (UID: \"31b5147c-2b03-40d2-bb73-4c0752279b1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.754392 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-cert\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.754722 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba-metrics-tls\") pod \"dns-default-gmdcv\" (UID: \"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba\") " pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.757166 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd864\" (UniqueName: \"kubernetes.io/projected/0ba3a48a-6733-41de-a3a5-fc6f00b429a2-kube-api-access-gd864\") pod \"package-server-manager-789f6589d5-hvhpp\" (UID: \"0ba3a48a-6733-41de-a3a5-fc6f00b429a2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.757313 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-certs\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.757918 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnmm9\" (UniqueName: \"kubernetes.io/projected/506d74f6-5ffd-4569-98db-949e699a9f22-kube-api-access-cnmm9\") pod \"ingress-operator-5b745b69d9-lb8k8\" (UID: \"506d74f6-5ffd-4569-98db-949e699a9f22\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.758068 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dslb\" (UniqueName: \"kubernetes.io/projected/2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3-kube-api-access-8dslb\") pod \"openshift-controller-manager-operator-756b6f6bc6-cxrvw\" (UID: \"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.758352 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8v7z\" (UniqueName: \"kubernetes.io/projected/d20ddb1e-869d-4625-a781-c2b35369fca5-kube-api-access-z8v7z\") pod \"apiserver-76f77b778f-h6x5b\" (UID: \"d20ddb1e-869d-4625-a781-c2b35369fca5\") " pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.760134 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59ckh\" (UniqueName: \"kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh\") pod \"marketplace-operator-79b997595-7rmh4\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.761733 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7k6x\" (UniqueName: \"kubernetes.io/projected/c0fb73bb-2bd2-460f-9d46-a1b44680e187-kube-api-access-b7k6x\") pod \"downloads-7954f5f757-2wj2q\" (UID: \"c0fb73bb-2bd2-460f-9d46-a1b44680e187\") " pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.765914 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkn6s\" (UniqueName: \"kubernetes.io/projected/20b6f5b4-21ff-412d-81aa-c29c4b4286f1-kube-api-access-qkn6s\") pod \"multus-admission-controller-857f4d67dd-ssxn4\" (UID: \"20b6f5b4-21ff-412d-81aa-c29c4b4286f1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.768013 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a91a0129-b4bd-4acd-bceb-3091d5714089-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.771011 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/668172a7-5728-423d-a32e-a0737d06dbe9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss5p6\" (UID: \"668172a7-5728-423d-a32e-a0737d06dbe9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.774039 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5fbc\" (UniqueName: \"kubernetes.io/projected/e3f1ef70-2f3f-472a-8261-4a3e6198b1df-kube-api-access-z5fbc\") pod \"ingress-canary-kwks8\" (UID: \"e3f1ef70-2f3f-472a-8261-4a3e6198b1df\") " pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.796823 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4prz\" (UniqueName: \"kubernetes.io/projected/c322b39b-fdab-4043-bfa5-3408acab0d10-kube-api-access-r4prz\") pod \"service-ca-operator-777779d784-w25jq\" (UID: \"c322b39b-fdab-4043-bfa5-3408acab0d10\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:31 crc kubenswrapper[4880]: I0126 15:56:31.797073 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdb9f\" (UniqueName: \"kubernetes.io/projected/f972ba6d-640d-462e-9dc0-f8ee5454691e-kube-api-access-fdb9f\") pod \"catalog-operator-68c6474976-sh2fp\" (UID: \"f972ba6d-640d-462e-9dc0-f8ee5454691e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.177164 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.180081 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:32 crc kubenswrapper[4880]: E0126 15:56:32.180804 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:32.680778213 +0000 UTC m=+118.166506930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.180992 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.182044 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.182402 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.187880 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.189702 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.189911 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.189998 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.191203 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.205541 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdw82\" (UniqueName: \"kubernetes.io/projected/c996fc87-b129-4efb-9355-f355726ea4a7-kube-api-access-pdw82\") pod \"migrator-59844c95c7-km4zv\" (UID: \"c996fc87-b129-4efb-9355-f355726ea4a7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.205797 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wcz6\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.207431 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffm7h\" (UniqueName: \"kubernetes.io/projected/adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa-kube-api-access-ffm7h\") pod \"machine-config-server-2296m\" (UID: \"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa\") " pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.210273 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzgv7\" (UniqueName: \"kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7\") pod \"collect-profiles-29490705-qzt6g\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.212201 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zxk\" (UniqueName: \"kubernetes.io/projected/a91a0129-b4bd-4acd-bceb-3091d5714089-kube-api-access-c6zxk\") pod \"control-plane-machine-set-operator-78cbb6b69f-54k89\" (UID: \"a91a0129-b4bd-4acd-bceb-3091d5714089\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.213318 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnmb\" (UniqueName: \"kubernetes.io/projected/cd4170db-b57f-4f72-a59b-242d9227f742-kube-api-access-jjnmb\") pod \"service-ca-9c57cc56f-vmcgp\" (UID: \"cd4170db-b57f-4f72-a59b-242d9227f742\") " pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.215360 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgwmb\" (UniqueName: \"kubernetes.io/projected/24d4b53c-609a-419f-803d-78e4b13b9afd-kube-api-access-wgwmb\") pod \"packageserver-d55dfcdfc-7dm6w\" (UID: \"24d4b53c-609a-419f-803d-78e4b13b9afd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.217785 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kflfw\" (UniqueName: \"kubernetes.io/projected/76a651d0-6847-4104-9478-f46f6d8d7c7a-kube-api-access-kflfw\") pod \"machine-config-operator-74547568cd-kzm8c\" (UID: \"76a651d0-6847-4104-9478-f46f6d8d7c7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.230242 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w6dz\" (UniqueName: \"kubernetes.io/projected/ff3638c1-682f-4bac-a083-ea8ecb9763a4-kube-api-access-5w6dz\") pod \"kube-storage-version-migrator-operator-b67b599dd-cfbdn\" (UID: \"ff3638c1-682f-4bac-a083-ea8ecb9763a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.230512 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-677lc\" (UniqueName: \"kubernetes.io/projected/c631c684-80bc-4f57-ae38-bb96c0db15d1-kube-api-access-677lc\") pod \"csi-hostpathplugin-pxxnr\" (UID: \"c631c684-80bc-4f57-ae38-bb96c0db15d1\") " pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.303713 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:32 crc kubenswrapper[4880]: E0126 15:56:32.304350 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:32.804331577 +0000 UTC m=+118.290060284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.407206 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:32 crc kubenswrapper[4880]: E0126 15:56:32.407544 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:32.907526493 +0000 UTC m=+118.393255200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.415877 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.434784 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.480162 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2296m" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.556464 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:32 crc kubenswrapper[4880]: E0126 15:56:32.556954 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.05692434 +0000 UTC m=+118.542653047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.816313 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" Jan 26 15:56:32 crc kubenswrapper[4880]: I0126 15:56:32.816822 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:32 crc kubenswrapper[4880]: E0126 15:56:32.839332 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.339300161 +0000 UTC m=+118.825028868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.171303 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.171854 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.67183567 +0000 UTC m=+119.157564377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.211259 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.220152 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-d4nqf" event={"ID":"9340b971-8db7-4e70-bbc9-0e4d7292b4e1","Type":"ContainerStarted","Data":"8fa29ad5a8a36dd4dcec3e2138f806ed322f0b9151a2cbb80cc496101ffefbae"} Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.222414 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" event={"ID":"e873659e-c74c-45e8-ba2b-13a16071c117","Type":"ContainerStarted","Data":"05841a050b6f4d177a9c1c5e2b8a8abad30d6bf53b7349e81d0db48025a8a4de"} Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.277274 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.281576 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.781465963 +0000 UTC m=+119.267194670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.282047 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.283061 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.783041722 +0000 UTC m=+119.268770429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.413975 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.414415 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:33.914390287 +0000 UTC m=+119.400118994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.516898 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.517408 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.017396088 +0000 UTC m=+119.503124795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.619695 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.620177 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.120149813 +0000 UTC m=+119.605878520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.620407 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.622541 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.122520742 +0000 UTC m=+119.608249449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.724563 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.724843 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.224798655 +0000 UTC m=+119.710527362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.725042 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.725584 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.225565683 +0000 UTC m=+119.711294390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.830093 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.830640 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.330622285 +0000 UTC m=+119.816350992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.832320 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.833980 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f9lnb"] Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.854107 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8"] Jan 26 15:56:33 crc kubenswrapper[4880]: W0126 15:56:33.934995 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode96b5221_3d1f_4f21_bb0f_4b01373ffdb0.slice/crio-84c507dce4a84ef638fcc97b43622d6e906f9f7265d724051d51aa00ed989564 WatchSource:0}: Error finding container 84c507dce4a84ef638fcc97b43622d6e906f9f7265d724051d51aa00ed989564: Status 404 returned error can't find the container with id 84c507dce4a84ef638fcc97b43622d6e906f9f7265d724051d51aa00ed989564 Jan 26 15:56:33 crc kubenswrapper[4880]: I0126 15:56:33.936450 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:33 crc kubenswrapper[4880]: E0126 15:56:33.936942 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.436925616 +0000 UTC m=+119.922654323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.115039 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.117133 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.617087936 +0000 UTC m=+120.102816643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.117360 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.117795 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.617784624 +0000 UTC m=+120.103513331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.218288 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.218885 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.718864118 +0000 UTC m=+120.204592825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.322461 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.323063 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.823042417 +0000 UTC m=+120.308771124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.364403 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" event={"ID":"125185c6-2155-4eef-ba91-e52ac6347421","Type":"ContainerStarted","Data":"0b09db30290395a7868f77de9356ac8bf83154a1b8331cd00f1c3844bb6025c3"} Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.364923 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2296m" event={"ID":"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa","Type":"ContainerStarted","Data":"0179bb594876178a813da02dcd98d270bff8c241403f79c759d0aac201b2a321"} Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.365056 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" event={"ID":"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0","Type":"ContainerStarted","Data":"84c507dce4a84ef638fcc97b43622d6e906f9f7265d724051d51aa00ed989564"} Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.365165 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" event={"ID":"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1","Type":"ContainerStarted","Data":"3ce449568866314292ed80893af097882c5fe210ff2af65880db69266dd00498"} Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.365260 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-q8tjj"] Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.365353 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54"] Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.365480 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.424751 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.425557 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:34.925532756 +0000 UTC m=+120.411261463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.527232 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.527994 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.027968504 +0000 UTC m=+120.513697281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: W0126 15:56:34.555763 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddab46cd9_c24c_481b_99ed_c3449c91fbb6.slice/crio-a909cea95feff39ef0548b2c5a1c7b20174cae2ff9e46feec40311b8262f13f4 WatchSource:0}: Error finding container a909cea95feff39ef0548b2c5a1c7b20174cae2ff9e46feec40311b8262f13f4: Status 404 returned error can't find the container with id a909cea95feff39ef0548b2c5a1c7b20174cae2ff9e46feec40311b8262f13f4 Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.628952 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.640100 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.13814767 +0000 UTC m=+120.623876377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.688077 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.733538 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.734159 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.23412895 +0000 UTC m=+120.719857657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.834772 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.835186 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.335144282 +0000 UTC m=+120.820873009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:34 crc kubenswrapper[4880]: I0126 15:56:34.942624 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:34 crc kubenswrapper[4880]: E0126 15:56:34.943271 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.443242898 +0000 UTC m=+120.928971605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.004816 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.044416 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.045327 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.545306266 +0000 UTC m=+121.031034973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.110057 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6ft2z"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.119206 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.145915 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.146355 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.646340949 +0000 UTC m=+121.132069656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: W0126 15:56:35.188356 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8272257_ba55_49eb_84c5_65cf2987c8ce.slice/crio-b89591e1a9c1304760412bfa4da62900521b8bc529df79dbd6345f0a441fe520 WatchSource:0}: Error finding container b89591e1a9c1304760412bfa4da62900521b8bc529df79dbd6345f0a441fe520: Status 404 returned error can't find the container with id b89591e1a9c1304760412bfa4da62900521b8bc529df79dbd6345f0a441fe520 Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.196253 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.236973 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qjrvl"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.246741 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.247149 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.747131216 +0000 UTC m=+121.232859923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.250374 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.251304 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.251496 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.255459 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.274725 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4vdpp"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.283677 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" event={"ID":"875a4230-6bbb-4b42-b959-33ceeef0e3c4","Type":"ContainerStarted","Data":"63a469ae102c82a2b87493b69b853c5821973a52de2fe56c84bccb545a2e1f3e"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.289447 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" event={"ID":"dab46cd9-c24c-481b-99ed-c3449c91fbb6","Type":"ContainerStarted","Data":"a909cea95feff39ef0548b2c5a1c7b20174cae2ff9e46feec40311b8262f13f4"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.290666 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.295886 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cg95b"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.302771 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.309575 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6ft2z" event={"ID":"c8272257-ba55-49eb-84c5-65cf2987c8ce","Type":"ContainerStarted","Data":"b89591e1a9c1304760412bfa4da62900521b8bc529df79dbd6345f0a441fe520"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.318839 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.323797 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.327015 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-d4nqf" event={"ID":"9340b971-8db7-4e70-bbc9-0e4d7292b4e1","Type":"ContainerStarted","Data":"f2c9db81d3f6eb2da2d6da4195bb2a8294c4b3826882dfd2b3279c7bc34b606d"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.346751 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-h6x5b"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.347399 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" event={"ID":"e873659e-c74c-45e8-ba2b-13a16071c117","Type":"ContainerStarted","Data":"5cc74c0b5c5a19b5349eeb954358f048dc78afd4b82ef55cd34839e5d5dadb92"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.348126 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.348578 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.848559249 +0000 UTC m=+121.334287966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.353046 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" event={"ID":"0ba3a48a-6733-41de-a3a5-fc6f00b429a2","Type":"ContainerStarted","Data":"5f7dbdadc00c8f6c31caf67453b0406416c06617ed2585c52c1c6787e37f0e85"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.354626 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" event={"ID":"3d13c678-e761-49ad-8fa9-390e3be41292","Type":"ContainerStarted","Data":"41cebaece111581b4306124a30c30474b2a7740e51454b2dcf66f9f5d7594c61"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.355737 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.363285 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wbm2m"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.365328 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2296m" event={"ID":"adb99a60-3ac7-4a9c-b8ba-c2bd6ac533aa","Type":"ContainerStarted","Data":"4f259e7a9369204dca7436e07aaaeecf0d4c6ce17b375b22a4b27e851444ada7"} Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.365629 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.367987 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.368179 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-d4nqf" podStartSLOduration=86.368158619 podStartE2EDuration="1m26.368158619s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:35.362052908 +0000 UTC m=+120.847781615" watchObservedRunningTime="2026-01-26 15:56:35.368158619 +0000 UTC m=+120.853887326" Jan 26 15:56:35 crc kubenswrapper[4880]: W0126 15:56:35.372837 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod506d74f6_5ffd_4569_98db_949e699a9f22.slice/crio-34e7df8b154d42ecfad25a4b54da81e111791bbde2b66965d74a049871ded436 WatchSource:0}: Error finding container 34e7df8b154d42ecfad25a4b54da81e111791bbde2b66965d74a049871ded436: Status 404 returned error can't find the container with id 34e7df8b154d42ecfad25a4b54da81e111791bbde2b66965d74a049871ded436 Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.383583 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-2wj2q"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.389177 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2296m" podStartSLOduration=7.389153162 podStartE2EDuration="7.389153162s" podCreationTimestamp="2026-01-26 15:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:35.388632929 +0000 UTC m=+120.874361646" watchObservedRunningTime="2026-01-26 15:56:35.389153162 +0000 UTC m=+120.874881869" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.449552 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.452065 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:35.951996971 +0000 UTC m=+121.437725688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.456514 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.482311 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.529282 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.564717 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.567172 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.067154569 +0000 UTC m=+121.552883276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.585207 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.585744 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.615248 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.653046 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.667047 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.668041 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.168015948 +0000 UTC m=+121.653744655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.709526 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.733960 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ssxn4"] Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.736095 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kwks8" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.769171 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.772169 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.272152996 +0000 UTC m=+121.757881703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.774107 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:35 crc kubenswrapper[4880]: W0126 15:56:35.800318 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20b6f5b4_21ff_412d_81aa_c29c4b4286f1.slice/crio-3f7dd0933987a29f2a6d23f16d19c1041a1a48266eb286a89fabe5ce2a6932b5 WatchSource:0}: Error finding container 3f7dd0933987a29f2a6d23f16d19c1041a1a48266eb286a89fabe5ce2a6932b5: Status 404 returned error can't find the container with id 3f7dd0933987a29f2a6d23f16d19c1041a1a48266eb286a89fabe5ce2a6932b5 Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.840945 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.879240 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.880371 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.380351464 +0000 UTC m=+121.866080171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:35 crc kubenswrapper[4880]: I0126 15:56:35.981840 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:35 crc kubenswrapper[4880]: E0126 15:56:35.982490 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.482398242 +0000 UTC m=+121.968127009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.054035 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.085193 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.085890 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.585868145 +0000 UTC m=+122.071596852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.387841 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.389255 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.395910 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:36.895833002 +0000 UTC m=+122.381561709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.570601 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.572877 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:36 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:36 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:36 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.572996 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.573093 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:37.071386469 +0000 UTC m=+122.557115176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.573179 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.574470 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:37.074434973 +0000 UTC m=+122.560163670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.833732 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.938702 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:37.43789606 +0000 UTC m=+122.923624767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:36 crc kubenswrapper[4880]: I0126 15:56:36.938959 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:36 crc kubenswrapper[4880]: E0126 15:56:36.942394 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:37.442297218 +0000 UTC m=+122.928025925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.046127 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.047031 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.047823 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:37.547790589 +0000 UTC m=+123.033519306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.115380 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 15:56:37 crc kubenswrapper[4880]: W0126 15:56:37.206908 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8af8728_cae3_44d4_9f71_79c69e8c5646.slice/crio-adeead3640a9f2ab879c4498b469638e1e6461d35b625afba096a6f059a25c16 WatchSource:0}: Error finding container adeead3640a9f2ab879c4498b469638e1e6461d35b625afba096a6f059a25c16: Status 404 returned error can't find the container with id adeead3640a9f2ab879c4498b469638e1e6461d35b625afba096a6f059a25c16 Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.244131 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" event={"ID":"d0fd7618-4cfa-41b4-b5fe-4e72e8a93df1","Type":"ContainerStarted","Data":"7bfdae5396a383c92f328fab35869d4e454c18bfca7c321c211423a67cdcac32"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.254789 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:37 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:37 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:37 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.254987 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.255538 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" event={"ID":"875a4230-6bbb-4b42-b959-33ceeef0e3c4","Type":"ContainerStarted","Data":"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.266501 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/216872c9-037e-486f-bf08-5e970a2bd5b1-metrics-certs\") pod \"network-metrics-daemon-pvdts\" (UID: \"216872c9-037e-486f-bf08-5e970a2bd5b1\") " pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.268358 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.569557 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.573813 4880 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4xvtl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.573875 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.574757 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.574694917 +0000 UTC m=+124.060423624 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.628924 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.629885 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.129864037 +0000 UTC m=+123.615592744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.644547 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.644919 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-pvdts" Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.707425 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" event={"ID":"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a","Type":"ContainerStarted","Data":"fb167ba06c2dfb0974cc3f1f08abafeb211f048ce54a51a769b6f8e2406be706"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.707512 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" event={"ID":"16aea370-960e-4bc7-a90c-fe7c7e9e6a4a","Type":"ContainerStarted","Data":"3af6bb534c06e271e5442ecc0313a54a578b18e2953008efad957584d3b5d527"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.709118 4880 csr.go:261] certificate signing request csr-82frf is approved, waiting to be issued Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.730346 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.734229 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.234172119 +0000 UTC m=+123.719900836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.756349 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" event={"ID":"55ae9b4d-9703-4852-8725-b8564ec87c82","Type":"ContainerStarted","Data":"c7649771cb92a7d6993cd28b60a2be44c6a617870d01fb20f87638e727573a0d"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.756472 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" event={"ID":"55ae9b4d-9703-4852-8725-b8564ec87c82","Type":"ContainerStarted","Data":"0dc6e1081739af98d70f7834d12c36a91c1dadc7b70c2d93b05bc43e0b1e1e28"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.766692 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" event={"ID":"d20ddb1e-869d-4625-a781-c2b35369fca5","Type":"ContainerStarted","Data":"6fbaf8245381275e1374e1bbeb57291c6b9a49cb50725de9b81a53e54cfeebab"} Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.857746 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.859648 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.35961848 +0000 UTC m=+123.845347277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.881233 4880 csr.go:257] certificate signing request csr-82frf is issued Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.987848 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:37 crc kubenswrapper[4880]: E0126 15:56:37.988388 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.488353831 +0000 UTC m=+123.974082568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:37 crc kubenswrapper[4880]: I0126 15:56:37.989831 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" event={"ID":"20b6f5b4-21ff-412d-81aa-c29c4b4286f1","Type":"ContainerStarted","Data":"3f7dd0933987a29f2a6d23f16d19c1041a1a48266eb286a89fabe5ce2a6932b5"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.073996 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" event={"ID":"668172a7-5728-423d-a32e-a0737d06dbe9","Type":"ContainerStarted","Data":"912471947afba833367f3546dba54a5802a57eafc78f1f8ae8053d9c9d4c8629"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.076419 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" event={"ID":"a91a0129-b4bd-4acd-bceb-3091d5714089","Type":"ContainerStarted","Data":"399360ecdff0794b3046dc072259e91bb6fa274697bfd929957c7f524150085c"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.078452 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" event={"ID":"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3","Type":"ContainerStarted","Data":"234650bfab935553fad4924b8e3b96e83413b13a14fb89edf84cae85039d39df"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.078492 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" event={"ID":"2e0e4f7c-0ad7-4b35-8e72-5cde9802c5c3","Type":"ContainerStarted","Data":"5e044a8c925641e67f9cea78a9ecaedb44768a18a1875b133e6c605afa6b785f"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.083626 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" event={"ID":"62b30286-ebfb-4bd1-a781-dfd2f6ab107e","Type":"ContainerStarted","Data":"e1f75cd42388f235eb43fa4de4c775ebf841d58344868f388ad596d266b97903"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.087911 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" event={"ID":"125185c6-2155-4eef-ba91-e52ac6347421","Type":"ContainerStarted","Data":"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.089281 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.091061 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.091562 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.591546147 +0000 UTC m=+124.077274854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.131302 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" event={"ID":"e1296f6c-3717-4a9c-abc1-d27c4b98dd10","Type":"ContainerStarted","Data":"0b5dc1fa530a01616d5f1ccf470a4dae1e7ee2978ac49ab7cf819bcbcc881575"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.131826 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" event={"ID":"e1296f6c-3717-4a9c-abc1-d27c4b98dd10","Type":"ContainerStarted","Data":"7598c9279427dee20a4c55730f36b7962487f114ebc7a5546a6f77b75672ee30"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.147151 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" event={"ID":"94edc90a-5195-4297-b015-feb66415fae5","Type":"ContainerStarted","Data":"540ebb347f2ff449159e8d6589643a6038b156ea453bab8671ba8d07d8c10eea"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.147232 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" event={"ID":"94edc90a-5195-4297-b015-feb66415fae5","Type":"ContainerStarted","Data":"234727691eb98a44801abadfd1e859dcd2e6c461f56dcdb1ee1ce4f648fdcb1a"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.148790 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.169929 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" event={"ID":"69b11523-e5c0-41b8-84d7-82cf9521626e","Type":"ContainerStarted","Data":"ea67cdaaaa5767ac14a3f0fc00b09a75e23cd1cf379b6decbf128e1b76096ac6"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.183480 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" event={"ID":"c996fc87-b129-4efb-9355-f355726ea4a7","Type":"ContainerStarted","Data":"7c4f4bd1ec74b1ae1d8dfb7d9fb0f051483d9a99334b52ab896b41d86b5541d6"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.183575 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" event={"ID":"c996fc87-b129-4efb-9355-f355726ea4a7","Type":"ContainerStarted","Data":"a8882b3e48dbd0da01f61bfb4d225a4c67d47d11e60fc31dd8b05eb436842d4f"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.189793 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:38 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:38 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:38 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.189862 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.192179 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.193602 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.693565764 +0000 UTC m=+124.179294471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.204258 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6ft2z" event={"ID":"c8272257-ba55-49eb-84c5-65cf2987c8ce","Type":"ContainerStarted","Data":"9612bc85dbe207b829a2a58e1cfe591cf1892289764de5dbc04ac55b035e0e6e"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.208106 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" event={"ID":"0d02c11b-b059-4d82-b864-5659fedb9fd5","Type":"ContainerStarted","Data":"3172857e23ed5b2476be953f0b998d8081605d7e9f69438cb8f96c1b307def40"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.210675 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" event={"ID":"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0","Type":"ContainerStarted","Data":"07769d95283f4beba83d7e0af406945c0d24eb11c0d3ee5a80ab554830d83d66"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.210717 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" event={"ID":"e96b5221-3d1f-4f21-bb0f-4b01373ffdb0","Type":"ContainerStarted","Data":"4f2288689d1734f6332dd09b736d8d472a512772c78bc758e62cee7924db9f13"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.214855 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2wj2q" event={"ID":"c0fb73bb-2bd2-460f-9d46-a1b44680e187","Type":"ContainerStarted","Data":"1cbf4a6aa46f267a377cde62deeabfbdb937fed19a6ffaa9201bf26a2eaacbf2"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.219428 4880 generic.go:334] "Generic (PLEG): container finished" podID="3d13c678-e761-49ad-8fa9-390e3be41292" containerID="cd8e52f682e2efa499fc492947b2cd473eb5b7cafac6302f0d83e448ce396caa" exitCode=0 Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.220906 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" event={"ID":"3d13c678-e761-49ad-8fa9-390e3be41292","Type":"ContainerDied","Data":"cd8e52f682e2efa499fc492947b2cd473eb5b7cafac6302f0d83e448ce396caa"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.233687 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.233822 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.491181 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.491653 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:38.991638329 +0000 UTC m=+124.477367036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.594039 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.596240 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.096167738 +0000 UTC m=+124.581896445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.598201 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.598286 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" event={"ID":"db5e6499-8551-4d89-a05e-936487c9aa56","Type":"ContainerStarted","Data":"fce493c8112a053855004c3dd882261ccb78088a711d7888d9992b04c20c03e8"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.600689 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" event={"ID":"140e65cc-7b24-4d1c-b558-07ca6f71a29c","Type":"ContainerStarted","Data":"5b8c616eb863ba2e9ded346e7670550d2315e07d62bd2bc4354d1a86ef850b35"} Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.711116 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.713225 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.213183373 +0000 UTC m=+124.698912140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.883953 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.884513 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-26 15:51:37 +0000 UTC, rotation deadline is 2026-10-16 15:51:17.479336789 +0000 UTC Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.884547 4880 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6311h54m38.594793363s for next certificate rotation Jan 26 15:56:38 crc kubenswrapper[4880]: E0126 15:56:38.884638 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.384623468 +0000 UTC m=+124.870352175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:38 crc kubenswrapper[4880]: I0126 15:56:38.940867 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" event={"ID":"e873659e-c74c-45e8-ba2b-13a16071c117","Type":"ContainerStarted","Data":"5e19a559d2f7e0c73731a22fd791aae6186502faa4f461e46e4b21929fd8268b"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.152670 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.153967 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.65393515 +0000 UTC m=+125.139663857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.273757 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.326633 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.826594316 +0000 UTC m=+125.312323023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.356059 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" event={"ID":"0ba3a48a-6733-41de-a3a5-fc6f00b429a2","Type":"ContainerStarted","Data":"7b6683994833dc939aba1d5cdbb6c81c0aec67a4146575dcdfc56b495361af27"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.378653 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.379419 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:39.879398129 +0000 UTC m=+125.365126836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.511928 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.512383 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.012351052 +0000 UTC m=+125.498079769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.521021 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" event={"ID":"dab46cd9-c24c-481b-99ed-c3449c91fbb6","Type":"ContainerStarted","Data":"57f70e95a790db4467bc1a1b640b64c84bf06157863f48c93508809ab2f98e23"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.524752 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:39 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:39 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:39 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.524929 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.533622 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" event={"ID":"506d74f6-5ffd-4569-98db-949e699a9f22","Type":"ContainerStarted","Data":"77e784a92136431e6cedbde8ec2fecfec075348f8832662a813f536e6de101e1"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.533686 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" event={"ID":"506d74f6-5ffd-4569-98db-949e699a9f22","Type":"ContainerStarted","Data":"34e7df8b154d42ecfad25a4b54da81e111791bbde2b66965d74a049871ded436"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.620715 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.628738 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.128688561 +0000 UTC m=+125.614417268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.658706 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" event={"ID":"920d5f22-6fa6-4af8-8a47-554a8cffab54","Type":"ContainerStarted","Data":"9bdf6ac9fc2c2e6e7be153d8827872e566a69e300ec7a8e2414c5331ddc5b80e"} Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.739198 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.739905 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.239877582 +0000 UTC m=+125.725606289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.844239 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.850889 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.350855178 +0000 UTC m=+125.836583885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:39 crc kubenswrapper[4880]: I0126 15:56:39.946065 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:39 crc kubenswrapper[4880]: E0126 15:56:39.946592 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.446567401 +0000 UTC m=+125.932296118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.098881 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:40 crc kubenswrapper[4880]: E0126 15:56:40.099423 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.599406352 +0000 UTC m=+126.085135059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.204163 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:40 crc kubenswrapper[4880]: E0126 15:56:40.205352 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.705330134 +0000 UTC m=+126.191058841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.345744 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:40 crc kubenswrapper[4880]: E0126 15:56:40.346367 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.846342402 +0000 UTC m=+126.332071109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.452987 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:40 crc kubenswrapper[4880]: E0126 15:56:40.453645 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:40.953623462 +0000 UTC m=+126.439352169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.882654 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:40 crc kubenswrapper[4880]: E0126 15:56:40.883326 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:41.383308025 +0000 UTC m=+126.869036742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.956853 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:40 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:40 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:40 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:40 crc kubenswrapper[4880]: I0126 15:56:40.957520 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.094701 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.095172 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:41.595128321 +0000 UTC m=+127.080857028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.102846 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" event={"ID":"db5e6499-8551-4d89-a05e-936487c9aa56","Type":"ContainerStarted","Data":"595dc71065a889070490104d69dc01f6e9cacface5de7a5982396c9adeb4892e"} Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.204663 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.206879 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:41.706838801 +0000 UTC m=+127.192567518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.325122 4880 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4xvtl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.325201 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.325947 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.326599 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:41.826577037 +0000 UTC m=+127.312305744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.435384 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.435930 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:41.935913738 +0000 UTC m=+127.421642445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.436397 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" event={"ID":"668172a7-5728-423d-a32e-a0737d06dbe9","Type":"ContainerStarted","Data":"9768ed3b756eb7ffaf81442db664ee4b249371ee90914c36b652866bfbaca04c"} Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.612154 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.613274 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.113197264 +0000 UTC m=+127.598925981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.745219 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" event={"ID":"140e65cc-7b24-4d1c-b558-07ca6f71a29c","Type":"ContainerStarted","Data":"2b93bd17bcabb6de4dbf70ac563365b046f5b37ca9982253a5aeb0c3dedfa535"} Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.746564 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:41 crc kubenswrapper[4880]: E0126 15:56:41.747256 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.247234345 +0000 UTC m=+127.732963052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.747613 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.747647 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:56:41 crc kubenswrapper[4880]: I0126 15:56:41.857457 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.072966 4880 patch_prober.go:28] interesting pod/console-f9d7485db-6ft2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.073464 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6ft2z" podUID="c8272257-ba55-49eb-84c5-65cf2987c8ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 26 15:56:42 crc kubenswrapper[4880]: E0126 15:56:42.074077 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.574031624 +0000 UTC m=+128.059760321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.094803 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:42 crc kubenswrapper[4880]: E0126 15:56:42.096360 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.596330829 +0000 UTC m=+128.082059536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.185199 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.202772 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:42 crc kubenswrapper[4880]: E0126 15:56:42.266751 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.76671595 +0000 UTC m=+128.252444667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.307631 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.308136 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.311659 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.312007 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.312111 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.312201 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.312838 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:42 crc kubenswrapper[4880]: E0126 15:56:42.313465 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.813423853 +0000 UTC m=+128.299152560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.426244 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:42 crc kubenswrapper[4880]: E0126 15:56:42.426884 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:42.926859019 +0000 UTC m=+128.412587726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:42 crc kubenswrapper[4880]: I0126 15:56:42.528176 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:43 crc kubenswrapper[4880]: E0126 15:56:43.189186 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:43.689142397 +0000 UTC m=+129.174871104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.201047 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:43 crc kubenswrapper[4880]: E0126 15:56:43.201949 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:43.70190608 +0000 UTC m=+129.187634787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.205800 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.205932 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.883798 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:43 crc kubenswrapper[4880]: E0126 15:56:43.884723 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:44.884691441 +0000 UTC m=+130.370420148 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.892562 4880 generic.go:334] "Generic (PLEG): container finished" podID="69b11523-e5c0-41b8-84d7-82cf9521626e" containerID="167f889c4796e4d73c097aeae92de453b2eeb32e262b31e1bc212471616ea456" exitCode=0 Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.915662 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.915896 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.952703 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.952804 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.960761 4880 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4xvtl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.960904 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 26 15:56:43 crc kubenswrapper[4880]: I0126 15:56:43.986032 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:43 crc kubenswrapper[4880]: E0126 15:56:43.987073 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:44.487048487 +0000 UTC m=+129.972777194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:44 crc kubenswrapper[4880]: I0126 15:56:44.066927 4880 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pxfwl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 26 15:56:44 crc kubenswrapper[4880]: I0126 15:56:44.067015 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 26 15:56:44 crc kubenswrapper[4880]: I0126 15:56:44.092102 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:44 crc kubenswrapper[4880]: E0126 15:56:44.097616 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:44.597576622 +0000 UTC m=+130.083305329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.266885 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.266961 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.270095 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:45 crc kubenswrapper[4880]: E0126 15:56:45.271144 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:46.271116126 +0000 UTC m=+131.756844873 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.271231 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:45 crc kubenswrapper[4880]: E0126 15:56:45.271807 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:45.771791293 +0000 UTC m=+131.257520000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.409709 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:45 crc kubenswrapper[4880]: E0126 15:56:45.410231 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:45.910211211 +0000 UTC m=+131.395939908 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:45 crc kubenswrapper[4880]: I0126 15:56:45.411144 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:45 crc kubenswrapper[4880]: E0126 15:56:45.413429 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:45.913408629 +0000 UTC m=+131.399137416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:45.664260 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:46 crc kubenswrapper[4880]: E0126 15:56:45.666001 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:46.165973166 +0000 UTC m=+131.651701873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.035949 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:46 crc kubenswrapper[4880]: E0126 15:56:46.036773 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:46.536754766 +0000 UTC m=+132.022483473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:46 crc kubenswrapper[4880]: E0126 15:56:46.038811 4880 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="3.772s" Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.038968 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.038987 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" event={"ID":"920d5f22-6fa6-4af8-8a47-554a8cffab54","Type":"ContainerStarted","Data":"e59e952662ca21a8eab2170f7082a53ec46cc8aa3240c8bc1899924a5cdf0b4b"} Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.039009 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" event={"ID":"62b30286-ebfb-4bd1-a781-dfd2f6ab107e","Type":"ContainerStarted","Data":"1905252de18c8dda2d49de01eabc84ecaff83663e33b3d0772a073f696729211"} Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.039872 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.039928 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.043567 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.049268 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.057157 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.137819 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:46 crc kubenswrapper[4880]: E0126 15:56:46.138490 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:46.638430615 +0000 UTC m=+132.124159322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.138713 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:46 crc kubenswrapper[4880]: E0126 15:56:46.139480 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:46.63942693 +0000 UTC m=+132.125155637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:46 crc kubenswrapper[4880]: I0126 15:56:46.162124 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" podStartSLOduration=96.161996832 podStartE2EDuration="1m36.161996832s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:46.159306196 +0000 UTC m=+131.645034913" watchObservedRunningTime="2026-01-26 15:56:46.161996832 +0000 UTC m=+131.647725539" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.095118 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:47 crc kubenswrapper[4880]: E0126 15:56:47.104150 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.104113522 +0000 UTC m=+133.589842229 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.114005 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.114085 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.164716 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-76rr8" podStartSLOduration=98.164685554 podStartE2EDuration="1m38.164685554s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:47.15840633 +0000 UTC m=+132.644135037" watchObservedRunningTime="2026-01-26 15:56:47.164685554 +0000 UTC m=+132.650414261" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.262058 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:47 crc kubenswrapper[4880]: E0126 15:56:47.262661 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:47.762635961 +0000 UTC m=+133.248364668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:47 crc kubenswrapper[4880]: E0126 15:56:47.608765 4880 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.565s" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.609070 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.609183 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-2wj2q" event={"ID":"c0fb73bb-2bd2-460f-9d46-a1b44680e187","Type":"ContainerStarted","Data":"67da49359038a07255c070ecf3baa0a88321bf464194c8106a0aae49c30147a3"} Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.615748 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:47 crc kubenswrapper[4880]: E0126 15:56:47.623236 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.123173246 +0000 UTC m=+133.608901953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.632047 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-46rpj" podStartSLOduration=98.631876319 podStartE2EDuration="1m38.631876319s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:47.608571128 +0000 UTC m=+133.094299835" watchObservedRunningTime="2026-01-26 15:56:47.631876319 +0000 UTC m=+133.117605036" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.648206 4880 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pxfwl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.648288 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.651992 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.652082 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.828302 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:47 crc kubenswrapper[4880]: E0126 15:56:47.830525 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.33050424 +0000 UTC m=+133.816232947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:47 crc kubenswrapper[4880]: I0126 15:56:47.844730 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6ft2z" podStartSLOduration=98.844691688 podStartE2EDuration="1m38.844691688s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:47.842367131 +0000 UTC m=+133.328095838" watchObservedRunningTime="2026-01-26 15:56:47.844691688 +0000 UTC m=+133.330420395" Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.176070 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:48 crc kubenswrapper[4880]: E0126 15:56:48.176670 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.676633896 +0000 UTC m=+134.162362603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.181626 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.181752 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.280384 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:48 crc kubenswrapper[4880]: E0126 15:56:48.282225 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.782199976 +0000 UTC m=+134.267928693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.458069 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:48 crc kubenswrapper[4880]: E0126 15:56:48.460505 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:48.96048622 +0000 UTC m=+134.446214927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.479477 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" podStartSLOduration=98.479427324 podStartE2EDuration="1m38.479427324s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:48.267770744 +0000 UTC m=+133.753499471" watchObservedRunningTime="2026-01-26 15:56:48.479427324 +0000 UTC m=+133.965156041" Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.481353 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-f9lnb" podStartSLOduration=98.48133314 podStartE2EDuration="1m38.48133314s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:48.477798154 +0000 UTC m=+133.963526851" watchObservedRunningTime="2026-01-26 15:56:48.48133314 +0000 UTC m=+133.967061847" Jan 26 15:56:48 crc kubenswrapper[4880]: I0126 15:56:48.563305 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:48 crc kubenswrapper[4880]: E0126 15:56:48.564105 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:49.064041125 +0000 UTC m=+134.549769852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:49.433487 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:50 crc kubenswrapper[4880]: E0126 15:56:49.433928 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:50.433906676 +0000 UTC m=+135.919635383 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:49.434601 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:49.434668 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.288037 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podStartSLOduration=101.287996875 podStartE2EDuration="1m41.287996875s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:49.435536406 +0000 UTC m=+134.921265123" watchObservedRunningTime="2026-01-26 15:56:50.287996875 +0000 UTC m=+135.773725582" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.323717 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:50 crc kubenswrapper[4880]: E0126 15:56:50.324689 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:50.824649745 +0000 UTC m=+136.310378452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.358228 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss5p6" podStartSLOduration=100.358199879 podStartE2EDuration="1m40.358199879s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:50.341840767 +0000 UTC m=+135.827569484" watchObservedRunningTime="2026-01-26 15:56:50.358199879 +0000 UTC m=+135.843928586" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.429778 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:50 crc kubenswrapper[4880]: E0126 15:56:50.430601 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:50.93056193 +0000 UTC m=+136.416290647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.443639 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.443744 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.452617 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-q8tjj" podStartSLOduration=101.452567499 podStartE2EDuration="1m41.452567499s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:50.448076189 +0000 UTC m=+135.933804906" watchObservedRunningTime="2026-01-26 15:56:50.452567499 +0000 UTC m=+135.938296236" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.550352 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:50 crc kubenswrapper[4880]: E0126 15:56:50.551802 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:51.051769207 +0000 UTC m=+136.537497914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.641354 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plrb8" podStartSLOduration=100.641310418 podStartE2EDuration="1m40.641310418s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:50.637095245 +0000 UTC m=+136.122823952" watchObservedRunningTime="2026-01-26 15:56:50.641310418 +0000 UTC m=+136.127039125" Jan 26 15:56:50 crc kubenswrapper[4880]: I0126 15:56:50.951860 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:50 crc kubenswrapper[4880]: E0126 15:56:50.952519 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:51.452490965 +0000 UTC m=+136.938219672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.146525 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:51 crc kubenswrapper[4880]: E0126 15:56:51.147053 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:51.647033966 +0000 UTC m=+137.132762673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.210838 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-2wj2q" podStartSLOduration=102.210816567 podStartE2EDuration="1m42.210816567s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:51.144555355 +0000 UTC m=+136.630284062" watchObservedRunningTime="2026-01-26 15:56:51.210816567 +0000 UTC m=+136.696545274" Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.217503 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" podStartSLOduration=101.21748096 podStartE2EDuration="1m41.21748096s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:51.209535616 +0000 UTC m=+136.695264333" watchObservedRunningTime="2026-01-26 15:56:51.21748096 +0000 UTC m=+136.703209677" Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.711278 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:51 crc kubenswrapper[4880]: E0126 15:56:51.712721 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.712695704 +0000 UTC m=+138.198424421 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.719526 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.719627 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.725992 4880 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pxfwl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.726113 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.730360 4880 patch_prober.go:28] interesting pod/console-f9d7485db-6ft2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.730401 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6ft2z" podUID="c8272257-ba55-49eb-84c5-65cf2987c8ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.812980 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:51 crc kubenswrapper[4880]: E0126 15:56:51.813611 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.31358814 +0000 UTC m=+137.799316847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:51 crc kubenswrapper[4880]: I0126 15:56:51.918920 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:51 crc kubenswrapper[4880]: E0126 15:56:51.927718 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.427686916 +0000 UTC m=+137.913415623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:51.960732 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cxrvw" podStartSLOduration=102.960705202 podStartE2EDuration="1m42.960705202s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:51.955940516 +0000 UTC m=+137.441669253" watchObservedRunningTime="2026-01-26 15:56:51.960705202 +0000 UTC m=+137.446433909" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:51.961492 4880 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.352s" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:51.961539 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" event={"ID":"69b11523-e5c0-41b8-84d7-82cf9521626e","Type":"ContainerDied","Data":"167f889c4796e4d73c097aeae92de453b2eeb32e262b31e1bc212471616ea456"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:51.961619 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:51.973352 4880 generic.go:334] "Generic (PLEG): container finished" podID="d20ddb1e-869d-4625-a781-c2b35369fca5" containerID="d6f5ff4139b1d4b77f4e202f092aef035bf9952e67b40f966c9a1b6cd38aa6af" exitCode=0 Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.068356 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.068897 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.568873609 +0000 UTC m=+138.054602316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.068977 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.070037 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.570020927 +0000 UTC m=+138.055749634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079629 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" event={"ID":"0d02c11b-b059-4d82-b864-5659fedb9fd5","Type":"ContainerStarted","Data":"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079688 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-54k89" event={"ID":"a91a0129-b4bd-4acd-bceb-3091d5714089","Type":"ContainerStarted","Data":"13eebbf2d9ec891b2d967e4268ea94ee8236ceb0d39ef98bc4462035ec191818"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079709 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" event={"ID":"c8af8728-cae3-44d4-9f71-79c69e8c5646","Type":"ContainerStarted","Data":"adeead3640a9f2ab879c4498b469638e1e6461d35b625afba096a6f059a25c16"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079724 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" event={"ID":"0ba3a48a-6733-41de-a3a5-fc6f00b429a2","Type":"ContainerStarted","Data":"bfe757dcbdf96343a76532ac68e4af3247ec00b91db40a8a79661ed8b582689d"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079744 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" event={"ID":"506d74f6-5ffd-4569-98db-949e699a9f22","Type":"ContainerStarted","Data":"ae11b3a7895285e88592724471389310eb35dd049699be192a86b1c564c506e1"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079774 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" event={"ID":"d20ddb1e-869d-4625-a781-c2b35369fca5","Type":"ContainerStarted","Data":"d6f5ff4139b1d4b77f4e202f092aef035bf9952e67b40f966c9a1b6cd38aa6af"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.079794 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" event={"ID":"d20ddb1e-869d-4625-a781-c2b35369fca5","Type":"ContainerDied","Data":"d6f5ff4139b1d4b77f4e202f092aef035bf9952e67b40f966c9a1b6cd38aa6af"} Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.081541 4880 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pxfwl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.081591 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.084072 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.174620 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.176410 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.676388051 +0000 UTC m=+138.162116758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.177323 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jtztn" podStartSLOduration=103.177294473 podStartE2EDuration="1m43.177294473s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:52.164120941 +0000 UTC m=+137.649849658" watchObservedRunningTime="2026-01-26 15:56:52.177294473 +0000 UTC m=+137.663023180" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.185679 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.185771 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.186137 4880 patch_prober.go:28] interesting pod/console-operator-58897d9998-4vdpp container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.186243 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podUID="94edc90a-5195-4297-b015-feb66415fae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/healthz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.193292 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.193388 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.193540 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.193608 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.207111 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qjrvl" podStartSLOduration=103.207080472 podStartE2EDuration="1m43.207080472s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:52.202163602 +0000 UTC m=+137.687892309" watchObservedRunningTime="2026-01-26 15:56:52.207080472 +0000 UTC m=+137.692809179" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.263038 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:52 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:52 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:52 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.263108 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.269544 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:52 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:52 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:52 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.269620 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.326768 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" podStartSLOduration=103.326745121 podStartE2EDuration="1m43.326745121s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:52.315098246 +0000 UTC m=+137.800826963" watchObservedRunningTime="2026-01-26 15:56:52.326745121 +0000 UTC m=+137.812473828" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.333374 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.336523 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.83650516 +0000 UTC m=+138.322233867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.459524 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.459960 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.959935681 +0000 UTC m=+138.445664398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.460151 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.460973 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:52.960960867 +0000 UTC m=+138.446689574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.562329 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.563278 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.06323821 +0000 UTC m=+138.548966927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.645124 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mkbhn" podStartSLOduration=102.645099283 podStartE2EDuration="1m42.645099283s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:52.643046424 +0000 UTC m=+138.128775131" watchObservedRunningTime="2026-01-26 15:56:52.645099283 +0000 UTC m=+138.130827990" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.665477 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.666085 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.166065187 +0000 UTC m=+138.651793894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.778752 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.779223 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.279200156 +0000 UTC m=+138.764928863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.889006 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" podStartSLOduration=103.888974823 podStartE2EDuration="1m43.888974823s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:52.758660223 +0000 UTC m=+138.244388940" watchObservedRunningTime="2026-01-26 15:56:52.888974823 +0000 UTC m=+138.374703530" Jan 26 15:56:52 crc kubenswrapper[4880]: I0126 15:56:52.892352 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:52 crc kubenswrapper[4880]: E0126 15:56:52.892828 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.392812777 +0000 UTC m=+138.878541484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:52.931140 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:52.942925 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.000814 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.001307 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.501284731 +0000 UTC m=+138.987013438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.034305 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" event={"ID":"20b6f5b4-21ff-412d-81aa-c29c4b4286f1","Type":"ContainerStarted","Data":"833fd90efe939b26e9cc3041305e0a5081ac1b8d33771ae0af387a4e7ffdeb6d"} Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.049797 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.064308 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" event={"ID":"c996fc87-b129-4efb-9355-f355726ea4a7","Type":"ContainerStarted","Data":"0d1e7e67ed40ae5b048fc65795686308626d5e61f66f534262e53c3e1e9ad93e"} Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.102599 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.107722 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.607703186 +0000 UTC m=+139.093431893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.123936 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" event={"ID":"920d5f22-6fa6-4af8-8a47-554a8cffab54","Type":"ContainerStarted","Data":"79031f3741da5b4a802f1bc7c5a27b9c7576c48042a6bbb7f08f6b3a94fc9dad"} Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.208468 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.208958 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.708937635 +0000 UTC m=+139.194666342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.222721 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:53 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:53 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:53 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.222787 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.235600 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.299150 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lb8k8" podStartSLOduration=103.299085481 podStartE2EDuration="1m43.299085481s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:53.27291376 +0000 UTC m=+138.758642477" watchObservedRunningTime="2026-01-26 15:56:53.299085481 +0000 UTC m=+138.784814188" Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.302872 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.314570 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.315914 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.815900052 +0000 UTC m=+139.301628759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.366316 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-km4zv" podStartSLOduration=103.366285186 podStartE2EDuration="1m43.366285186s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:53.356134967 +0000 UTC m=+138.841863674" watchObservedRunningTime="2026-01-26 15:56:53.366285186 +0000 UTC m=+138.852013893" Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.421912 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.422360 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:53.922339707 +0000 UTC m=+139.408068414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.425452 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-j9b2x" podStartSLOduration=103.425395432 podStartE2EDuration="1m43.425395432s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:53.421021875 +0000 UTC m=+138.906750582" watchObservedRunningTime="2026-01-26 15:56:53.425395432 +0000 UTC m=+138.911124139" Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.489018 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-w25jq"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.529086 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.529516 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.029503691 +0000 UTC m=+139.515232398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.543424 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-gmdcv"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.574930 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.702683 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.706088 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.206058672 +0000 UTC m=+139.691787379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.737565 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kwks8"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.808634 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.809571 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.309546675 +0000 UTC m=+139.795275382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.858760 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vmcgp"] Jan 26 15:56:53 crc kubenswrapper[4880]: I0126 15:56:53.914134 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:53 crc kubenswrapper[4880]: E0126 15:56:53.914810 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.414788311 +0000 UTC m=+139.900517028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:53 crc kubenswrapper[4880]: W0126 15:56:53.951847 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3f1ef70_2f3f_472a_8261_4a3e6198b1df.slice/crio-c5551b0daccc3a347822000f05b082d9b2649f9bb1beb7d3bc4334744e282d92 WatchSource:0}: Error finding container c5551b0daccc3a347822000f05b082d9b2649f9bb1beb7d3bc4334744e282d92: Status 404 returned error can't find the container with id c5551b0daccc3a347822000f05b082d9b2649f9bb1beb7d3bc4334744e282d92 Jan 26 15:56:53 crc kubenswrapper[4880]: W0126 15:56:53.953514 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5dee4c2_1f37_42e2_b3f3_5edf6a6233ba.slice/crio-9ae4140cb6a91e04d4efb848c8fa32b03b3aa83ee43dfb5ff502668732453227 WatchSource:0}: Error finding container 9ae4140cb6a91e04d4efb848c8fa32b03b3aa83ee43dfb5ff502668732453227: Status 404 returned error can't find the container with id 9ae4140cb6a91e04d4efb848c8fa32b03b3aa83ee43dfb5ff502668732453227 Jan 26 15:56:54 crc kubenswrapper[4880]: W0126 15:56:54.014529 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd4170db_b57f_4f72_a59b_242d9227f742.slice/crio-3dc01354bae15e5dbfb96742c2328fca29b71dcafce099b65d4df14ddd916108 WatchSource:0}: Error finding container 3dc01354bae15e5dbfb96742c2328fca29b71dcafce099b65d4df14ddd916108: Status 404 returned error can't find the container with id 3dc01354bae15e5dbfb96742c2328fca29b71dcafce099b65d4df14ddd916108 Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.016477 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.017127 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.517110766 +0000 UTC m=+140.002839473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.047553 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pxxnr"] Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.119040 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.119972 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.619947542 +0000 UTC m=+140.105676249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.145271 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-pvdts"] Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.206824 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:54 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:54 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:54 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.206898 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.222143 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.222655 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.722637416 +0000 UTC m=+140.208366123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.261997 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" event={"ID":"d20ddb1e-869d-4625-a781-c2b35369fca5","Type":"ContainerStarted","Data":"dda29ef2d41058969077b96dae3c5943c181cc5f842c02aedf27477d1b9023ae"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.273880 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kwks8" event={"ID":"e3f1ef70-2f3f-472a-8261-4a3e6198b1df","Type":"ContainerStarted","Data":"c5551b0daccc3a347822000f05b082d9b2649f9bb1beb7d3bc4334744e282d92"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.315100 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" event={"ID":"69b11523-e5c0-41b8-84d7-82cf9521626e","Type":"ContainerStarted","Data":"a43fcad96800d315bb6afd087111410bc1a6f85f410e2f363411914b743cc6ef"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.316364 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.337630 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.337730 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.837707883 +0000 UTC m=+140.323436590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.338102 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.338913 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.838892952 +0000 UTC m=+140.324621729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.352287 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" event={"ID":"24d4b53c-609a-419f-803d-78e4b13b9afd","Type":"ContainerStarted","Data":"8fd652b0e9d8f1cb23a511c8f929ea6bfdb392217f2f82e86fa9e30890a36dec"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.386082 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" event={"ID":"3d13c678-e761-49ad-8fa9-390e3be41292","Type":"ContainerStarted","Data":"df39c851820b21f88eb522c42c726eff40577801791b66ad14aaeb8fa128e420"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.421003 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gmdcv" event={"ID":"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba","Type":"ContainerStarted","Data":"9ae4140cb6a91e04d4efb848c8fa32b03b3aa83ee43dfb5ff502668732453227"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.445187 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.446164 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:54.946138956 +0000 UTC m=+140.431867663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.465354 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" event={"ID":"c322b39b-fdab-4043-bfa5-3408acab0d10","Type":"ContainerStarted","Data":"e88bb8f2c0c47b4e39181fc409e7d6c402f864dcaf2d0621d433bf2fbd13c92b"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.483328 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" podStartSLOduration=105.483290916 podStartE2EDuration="1m45.483290916s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:54.480991669 +0000 UTC m=+139.966720376" watchObservedRunningTime="2026-01-26 15:56:54.483290916 +0000 UTC m=+139.969019623" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.488931 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" event={"ID":"76a651d0-6847-4104-9478-f46f6d8d7c7a","Type":"ContainerStarted","Data":"ed65d54f6c7b7200f394c19e3d19261dc2c139bde733f51a623b848217e5f451"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.534719 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" event={"ID":"ff3638c1-682f-4bac-a083-ea8ecb9763a4","Type":"ContainerStarted","Data":"bf4d8b1939097175f1e87c800651f49d53fd66d17b17f2a5e936044f90d89920"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.544471 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" event={"ID":"62b30286-ebfb-4bd1-a781-dfd2f6ab107e","Type":"ContainerStarted","Data":"e8351f01dd269fb6b3d19f47666c0e8cb1be4c8a695cc1fd169ae2c10cc342b8"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.550130 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.551530 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.051511986 +0000 UTC m=+140.537240693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.555891 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" event={"ID":"c8af8728-cae3-44d4-9f71-79c69e8c5646","Type":"ContainerStarted","Data":"bb2d56aac038c8b94316087d7e36c62df80a79abf07acb027f55b010328daf6e"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.557269 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.576666 4880 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7rmh4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.576740 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.577264 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" event={"ID":"55ae9b4d-9703-4852-8725-b8564ec87c82","Type":"ContainerStarted","Data":"bdc16325b471578f4d2a59abca61c65cb9f00414925c3ca57d028d7c488259d1"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.582048 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" podStartSLOduration=104.582013572 podStartE2EDuration="1m44.582013572s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:54.572173541 +0000 UTC m=+140.057902258" watchObservedRunningTime="2026-01-26 15:56:54.582013572 +0000 UTC m=+140.067742279" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.600883 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" event={"ID":"31b5147c-2b03-40d2-bb73-4c0752279b1c","Type":"ContainerStarted","Data":"9122e95f7c91a7562f5f8277be6fb95efb5e550a5dd58dff5d1a42164b57868a"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.625644 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" event={"ID":"f972ba6d-640d-462e-9dc0-f8ee5454691e","Type":"ContainerStarted","Data":"6fa8ed4e28fa39334252430c869fba5f76b8cefbc0ace4fecb26e754d9fbb1d7"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.650492 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8kg54" podStartSLOduration=105.650472038 podStartE2EDuration="1m45.650472038s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:54.649871393 +0000 UTC m=+140.135600100" watchObservedRunningTime="2026-01-26 15:56:54.650472038 +0000 UTC m=+140.136200745" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.651070 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podStartSLOduration=104.651063103 podStartE2EDuration="1m44.651063103s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:54.616939017 +0000 UTC m=+140.102667724" watchObservedRunningTime="2026-01-26 15:56:54.651063103 +0000 UTC m=+140.136791810" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.651102 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.653081 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.153061501 +0000 UTC m=+140.638790208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.657416 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" event={"ID":"cd4170db-b57f-4f72-a59b-242d9227f742","Type":"ContainerStarted","Data":"3dc01354bae15e5dbfb96742c2328fca29b71dcafce099b65d4df14ddd916108"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.675752 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" event={"ID":"de011de4-f35b-4f67-b436-c965f427dbf2","Type":"ContainerStarted","Data":"fcc2519ad75d0cc1a211b6f8e81220dd79383c619a8fbdf40c464da938e53ea9"} Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.681997 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-wbm2m" podStartSLOduration=105.681975099 podStartE2EDuration="1m45.681975099s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:54.68078703 +0000 UTC m=+140.166515737" watchObservedRunningTime="2026-01-26 15:56:54.681975099 +0000 UTC m=+140.167703806" Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.778146 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.778714 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.278697276 +0000 UTC m=+140.764425983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.880248 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.880694 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.380673633 +0000 UTC m=+140.866402340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:54 crc kubenswrapper[4880]: I0126 15:56:54.982741 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:54 crc kubenswrapper[4880]: E0126 15:56:54.983231 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.483210912 +0000 UTC m=+140.968939619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.083990 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.084544 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.58447455 +0000 UTC m=+141.070203267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.189258 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.189853 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.68983904 +0000 UTC m=+141.175567747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.194861 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:55 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:55 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:55 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.194935 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.256584 4880 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7rmh4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.256938 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.256782 4880 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7rmh4 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.257004 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.295833 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.296301 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.796278775 +0000 UTC m=+141.282007482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.397596 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.398178 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.898162258 +0000 UTC m=+141.383890965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.484481 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.485124 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.498704 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.499454 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:55.999388567 +0000 UTC m=+141.485117274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.600688 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.601273 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.101254779 +0000 UTC m=+141.586983486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.701778 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.701918 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.201884992 +0000 UTC m=+141.687613709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.702607 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.702974 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.202963369 +0000 UTC m=+141.688692076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.710540 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" event={"ID":"31b5147c-2b03-40d2-bb73-4c0752279b1c","Type":"ContainerStarted","Data":"811a2d3d283bd019721736e7fc8fe838a1d906d6dac06cc85035e99eaf096d35"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.711775 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.713280 4880 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-52d79 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.713346 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" podUID="31b5147c-2b03-40d2-bb73-4c0752279b1c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.722991 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" event={"ID":"f972ba6d-640d-462e-9dc0-f8ee5454691e","Type":"ContainerStarted","Data":"92a7d6b4d61229cc1df8bb0b8944caa789bd626c8fff0333376a1e37804444d6"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.724313 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.726356 4880 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-sh2fp container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.726418 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" podUID="f972ba6d-640d-462e-9dc0-f8ee5454691e" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.735332 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" event={"ID":"cd4170db-b57f-4f72-a59b-242d9227f742","Type":"ContainerStarted","Data":"66f73bce11f906f380acaefd39db86c17ec0df162461964809218dc894957600"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.745931 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" podStartSLOduration=105.74590982 podStartE2EDuration="1m45.74590982s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:55.743275556 +0000 UTC m=+141.229004273" watchObservedRunningTime="2026-01-26 15:56:55.74590982 +0000 UTC m=+141.231638527" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.749370 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" event={"ID":"ff3638c1-682f-4bac-a083-ea8ecb9763a4","Type":"ContainerStarted","Data":"2954f9be143d8dc5cd051b8b36382dbfa59b6937664f23fce212d151d20c80ab"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.777923 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" event={"ID":"24d4b53c-609a-419f-803d-78e4b13b9afd","Type":"ContainerStarted","Data":"7e6479e02bf9345e2f44e6ef11fa4beecfb8c2917da7dd0f52ba28a586702eaa"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.778781 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.781229 4880 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7dm6w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.781295 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" podUID="24d4b53c-609a-419f-803d-78e4b13b9afd" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.803145 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvdts" event={"ID":"216872c9-037e-486f-bf08-5e970a2bd5b1","Type":"ContainerStarted","Data":"b98004b2d5db93edbe2c38f695f95699d169dcacd836f561d89781e6d28a0225"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.803890 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vmcgp" podStartSLOduration=105.803875009 podStartE2EDuration="1m45.803875009s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:55.799529403 +0000 UTC m=+141.285258110" watchObservedRunningTime="2026-01-26 15:56:55.803875009 +0000 UTC m=+141.289603716" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.804134 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.806261 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.306227687 +0000 UTC m=+141.791956434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.835067 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" event={"ID":"de011de4-f35b-4f67-b436-c965f427dbf2","Type":"ContainerStarted","Data":"e2c500cf641a40de7423e5cf47deddb768e46931817340b4bde56a0b79906e5f"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.873979 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kwks8" event={"ID":"e3f1ef70-2f3f-472a-8261-4a3e6198b1df","Type":"ContainerStarted","Data":"e3269c0b1e70bd747d13c76739f33f71c4820487791cdde9692d3f7dfcd74900"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.891800 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" event={"ID":"c631c684-80bc-4f57-ae38-bb96c0db15d1","Type":"ContainerStarted","Data":"1debe19ddb57bd527a321783b549324332d072427cbafe61760578ac8159ed30"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.914332 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:55 crc kubenswrapper[4880]: E0126 15:56:55.914791 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.414778683 +0000 UTC m=+141.900507390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.916281 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-cfbdn" podStartSLOduration=105.91626427 podStartE2EDuration="1m45.91626427s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:55.90197831 +0000 UTC m=+141.387707027" watchObservedRunningTime="2026-01-26 15:56:55.91626427 +0000 UTC m=+141.401992977" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.918662 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" podStartSLOduration=105.918643988 podStartE2EDuration="1m45.918643988s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:55.875549733 +0000 UTC m=+141.361278440" watchObservedRunningTime="2026-01-26 15:56:55.918643988 +0000 UTC m=+141.404372695" Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.928828 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gmdcv" event={"ID":"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba","Type":"ContainerStarted","Data":"1e60170ab25339dcba838bef9edfa64c895f5ff720563b08c6d35515e82f3dcb"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.976657 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" event={"ID":"20b6f5b4-21ff-412d-81aa-c29c4b4286f1","Type":"ContainerStarted","Data":"7a8a5e22673ceaaa3fbe7642477b2a40c6f2c0545012dc12d314fe222f5feb24"} Jan 26 15:56:55 crc kubenswrapper[4880]: I0126 15:56:55.978783 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" podStartSLOduration=106.97876377 podStartE2EDuration="1m46.97876377s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:55.976684129 +0000 UTC m=+141.462412826" watchObservedRunningTime="2026-01-26 15:56:55.97876377 +0000 UTC m=+141.464492477" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.015077 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.016352 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.516329139 +0000 UTC m=+142.002057856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.036303 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" event={"ID":"c322b39b-fdab-4043-bfa5-3408acab0d10","Type":"ContainerStarted","Data":"aea81e1a4ce9d656f299a284290302f258913c604afb69de6cc92097dcdfe1e2"} Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.052201 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" podStartSLOduration=106.052171966 podStartE2EDuration="1m46.052171966s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:56.049531292 +0000 UTC m=+141.535260009" watchObservedRunningTime="2026-01-26 15:56:56.052171966 +0000 UTC m=+141.537900673" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.056528 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" event={"ID":"76a651d0-6847-4104-9478-f46f6d8d7c7a","Type":"ContainerStarted","Data":"09aa0c3040f844f1107ffc5472f545459c49d54e86359dc4bd003863d45dce83"} Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.058616 4880 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7rmh4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.058678 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.067585 4880 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-cg95b container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.067994 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" podUID="69b11523-e5c0-41b8-84d7-82cf9521626e" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.120788 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.127627 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.627604592 +0000 UTC m=+142.113333389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.146333 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kwks8" podStartSLOduration=28.146307291 podStartE2EDuration="28.146307291s" podCreationTimestamp="2026-01-26 15:56:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:56.146002713 +0000 UTC m=+141.631731420" watchObservedRunningTime="2026-01-26 15:56:56.146307291 +0000 UTC m=+141.632035998" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.194411 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:56 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:56 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:56 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.194547 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.221226 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.221689 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.721667185 +0000 UTC m=+142.207395902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.232014 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.290137 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" podStartSLOduration=106.29011022 podStartE2EDuration="1m46.29011022s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:56.29011114 +0000 UTC m=+141.775839847" watchObservedRunningTime="2026-01-26 15:56:56.29011022 +0000 UTC m=+141.775838927" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.322671 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.323185 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.823167249 +0000 UTC m=+142.308895956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.424887 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.425283 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:56.925261238 +0000 UTC m=+142.410989945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.485827 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ssxn4" podStartSLOduration=106.48580254 podStartE2EDuration="1m46.48580254s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:56.479557348 +0000 UTC m=+141.965286065" watchObservedRunningTime="2026-01-26 15:56:56.48580254 +0000 UTC m=+141.971531247" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.488268 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-w25jq" podStartSLOduration=106.48825615 podStartE2EDuration="1m46.48825615s" podCreationTimestamp="2026-01-26 15:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:56.38160869 +0000 UTC m=+141.867337417" watchObservedRunningTime="2026-01-26 15:56:56.48825615 +0000 UTC m=+141.973984857" Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.525933 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.526659 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.02664597 +0000 UTC m=+142.512374677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.628221 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.628755 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.128731198 +0000 UTC m=+142.614459905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.729753 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.730164 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.230147691 +0000 UTC m=+142.715876398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.831770 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.832262 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.33222715 +0000 UTC m=+142.817955877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.832381 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.832881 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.332867765 +0000 UTC m=+142.818596472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:56 crc kubenswrapper[4880]: I0126 15:56:56.933919 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:56 crc kubenswrapper[4880]: E0126 15:56:56.934497 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.434462352 +0000 UTC m=+142.920191049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.035520 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.036019 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.535996567 +0000 UTC m=+143.021725264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.070520 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-gmdcv" event={"ID":"e5dee4c2-1f37-42e2-b3f3-5edf6a6233ba","Type":"ContainerStarted","Data":"2da91346fa103e103d75bf29fda6700adb33f9aaee323a3c59de63e32b2dfd00"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.077107 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvdts" event={"ID":"216872c9-037e-486f-bf08-5e970a2bd5b1","Type":"ContainerStarted","Data":"58c1100651c41f11989db934e23f9fcbd824af29ce14c039c7a979497a47c0e2"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.077189 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-pvdts" event={"ID":"216872c9-037e-486f-bf08-5e970a2bd5b1","Type":"ContainerStarted","Data":"71043664d9770f764275011ac3e834571232916d50ca15cf2b72007d4a9ca8fe"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.091304 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kzm8c" event={"ID":"76a651d0-6847-4104-9478-f46f6d8d7c7a","Type":"ContainerStarted","Data":"777c2980e031ae5101025e3004b20dda57a3bca5ad8de40c4cab7eb4bf061dfe"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.111835 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.112908 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.116877 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" event={"ID":"d20ddb1e-869d-4625-a781-c2b35369fca5","Type":"ContainerStarted","Data":"61810ad0ac30838719a00b1384389f8d94b35705ab11b1ce5dc45aff6ee4a97e"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.124619 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.124996 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.134988 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" event={"ID":"c631c684-80bc-4f57-ae38-bb96c0db15d1","Type":"ContainerStarted","Data":"c4978b12e7e86d01c2b76cb7f5704e5efd68c6205967460e0beb61fce4237a55"} Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.136330 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.138302 4880 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7rmh4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.138390 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.138998 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.638961158 +0000 UTC m=+143.124689885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.144998 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.145126 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.145193 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.152647 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-52d79" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.154362 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.654342394 +0000 UTC m=+143.140071101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.160633 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-szk2t" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.164060 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-sh2fp" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.164878 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-gmdcv" podStartSLOduration=28.164852051 podStartE2EDuration="28.164852051s" podCreationTimestamp="2026-01-26 15:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:57.131297749 +0000 UTC m=+142.617026456" watchObservedRunningTime="2026-01-26 15:56:57.164852051 +0000 UTC m=+142.650580758" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.181316 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.188925 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:57 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:57 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:57 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.188993 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.195017 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.195517 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.245272 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" podStartSLOduration=108.245249319 podStartE2EDuration="1m48.245249319s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:57.195273046 +0000 UTC m=+142.681001763" watchObservedRunningTime="2026-01-26 15:56:57.245249319 +0000 UTC m=+142.730978026" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.248164 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.248402 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.249073 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.249849 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.749825171 +0000 UTC m=+143.235553878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.269684 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.322216 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-pvdts" podStartSLOduration=108.322176012 podStartE2EDuration="1m48.322176012s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:57.306308593 +0000 UTC m=+142.792037300" watchObservedRunningTime="2026-01-26 15:56:57.322176012 +0000 UTC m=+142.807904719" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.329601 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.358714 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.359176 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.859161427 +0000 UTC m=+143.344890134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.439813 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.462048 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.462658 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:57.96263294 +0000 UTC m=+143.448361647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.563756 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.564343 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.064325808 +0000 UTC m=+143.550054515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.667329 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.667978 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.167938254 +0000 UTC m=+143.653666971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.769707 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.770525 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.270506625 +0000 UTC m=+143.756235332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.876264 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.876786 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.376762356 +0000 UTC m=+143.862491063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.935036 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7dm6w" Jan 26 15:56:57 crc kubenswrapper[4880]: I0126 15:56:57.982095 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:57 crc kubenswrapper[4880]: E0126 15:56:57.982599 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.482581606 +0000 UTC m=+143.968310313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.084201 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.084832 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.584809528 +0000 UTC m=+144.070538235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.122119 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.169484 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"35862492-137b-47ad-9547-9e54cb19674c","Type":"ContainerStarted","Data":"ec5a093822e4bb60b6c4c8ffd6869fd79f9bd43992893c9b8dbef04f510c765c"} Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.171689 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-gmdcv" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.201320 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.202015 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.701999327 +0000 UTC m=+144.187728024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.207140 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:58 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:58 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:58 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.207220 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.303108 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.303413 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.803352757 +0000 UTC m=+144.289081474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.304257 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.306135 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.806110485 +0000 UTC m=+144.291839302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.405269 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.405948 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:58.905924878 +0000 UTC m=+144.391653585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.439912 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.441300 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.448680 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.461183 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.508060 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.508626 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.008607561 +0000 UTC m=+144.494336268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.591383 4880 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-cg95b container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]log ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/max-in-flight-filter ok Jan 26 15:56:58 crc kubenswrapper[4880]: [-]poststarthook/storage-object-count-tracker-hook failed: reason withheld Jan 26 15:56:58 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.591489 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" podUID="69b11523-e5c0-41b8-84d7-82cf9521626e" containerName="openshift-config-operator" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.608854 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.609225 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqfwj\" (UniqueName: \"kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.609271 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.609301 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.609669 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.109627994 +0000 UTC m=+144.595356711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.640377 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.641770 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.644552 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.691581 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716062 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqfwj\" (UniqueName: \"kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716132 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716169 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716214 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716287 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716328 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.716360 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmdfr\" (UniqueName: \"kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.717031 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.217008422 +0000 UTC m=+144.702737179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.717818 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.727057 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.759908 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqfwj\" (UniqueName: \"kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj\") pod \"certified-operators-5d9pw\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.770008 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.817698 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.817977 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.818014 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmdfr\" (UniqueName: \"kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.818110 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.318063496 +0000 UTC m=+144.803792213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.818493 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.818877 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.818990 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.857279 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.858827 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.866335 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmdfr\" (UniqueName: \"kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr\") pod \"community-operators-4xxrk\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.896820 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.900177 4880 patch_prober.go:28] interesting pod/apiserver-76f77b778f-h6x5b container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]log ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]etcd ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/max-in-flight-filter ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 26 15:56:58 crc kubenswrapper[4880]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 26 15:56:58 crc kubenswrapper[4880]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/project.openshift.io-projectcache ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-startinformers ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 26 15:56:58 crc kubenswrapper[4880]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 26 15:56:58 crc kubenswrapper[4880]: livez check failed Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.900275 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" podUID="d20ddb1e-869d-4625-a781-c2b35369fca5" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.920605 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.920681 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.920735 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvxx6\" (UniqueName: \"kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.920772 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:58 crc kubenswrapper[4880]: E0126 15:56:58.921172 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.421154729 +0000 UTC m=+144.906883436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:58 crc kubenswrapper[4880]: I0126 15:56:58.986555 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.028377 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.028653 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvxx6\" (UniqueName: \"kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.028814 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.528776303 +0000 UTC m=+145.014505020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.028954 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.029024 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.029703 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.029745 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.061609 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.064052 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.068170 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.083106 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvxx6\" (UniqueName: \"kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6\") pod \"certified-operators-r6zxd\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.144203 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjdcs\" (UniqueName: \"kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.144285 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.144330 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.144351 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.144704 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.64469104 +0000 UTC m=+145.130419747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.183550 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" event={"ID":"c631c684-80bc-4f57-ae38-bb96c0db15d1","Type":"ContainerStarted","Data":"ed33694c60ee9241d149c5233ec8539b8e5ceade33cc0dbebcdaa2fbec95a636"} Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.186951 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"35862492-137b-47ad-9547-9e54cb19674c","Type":"ContainerStarted","Data":"e802536ad089063a58c1ec07c30c43a8d11c20f2b172048560c4d2c10849d7f4"} Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.187341 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.197325 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:56:59 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:56:59 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:56:59 crc kubenswrapper[4880]: healthz check failed Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.197477 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.207118 4880 generic.go:334] "Generic (PLEG): container finished" podID="de011de4-f35b-4f67-b436-c965f427dbf2" containerID="e2c500cf641a40de7423e5cf47deddb768e46931817340b4bde56a0b79906e5f" exitCode=0 Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.207338 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" event={"ID":"de011de4-f35b-4f67-b436-c965f427dbf2","Type":"ContainerDied","Data":"e2c500cf641a40de7423e5cf47deddb768e46931817340b4bde56a0b79906e5f"} Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.228156 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.228124562 podStartE2EDuration="2.228124562s" podCreationTimestamp="2026-01-26 15:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:56:59.214301694 +0000 UTC m=+144.700030401" watchObservedRunningTime="2026-01-26 15:56:59.228124562 +0000 UTC m=+144.713853269" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.247821 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.248701 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjdcs\" (UniqueName: \"kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.249132 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.249164 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.250224 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.750199593 +0000 UTC m=+145.235928300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.254403 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.255843 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.288094 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjdcs\" (UniqueName: \"kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs\") pod \"community-operators-t9qds\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.288367 4880 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.324482 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.350004 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.350455 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.850425565 +0000 UTC m=+145.336154272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.429747 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.454177 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.454754 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 15:56:59.954733379 +0000 UTC m=+145.440462086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.556725 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: E0126 15:56:59.557408 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 15:57:00.057394011 +0000 UTC m=+145.543122718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zc2dp" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.570606 4880 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-26T15:56:59.288415058Z","Handler":null,"Name":""} Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.579514 4880 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.579571 4880 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.581138 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.658670 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.664234 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.672147 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.759961 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.764224 4880 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.764267 4880 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.801354 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.806970 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zc2dp\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:56:59 crc kubenswrapper[4880]: W0126 15:56:59.812027 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f25591a_acee_4cc9_98b5_ca45d47ef1bf.slice/crio-5c8f20307c46ae6432f5adebbfa31b9b80e07a56e43541b298ce066b4f93197a WatchSource:0}: Error finding container 5c8f20307c46ae6432f5adebbfa31b9b80e07a56e43541b298ce066b4f93197a: Status 404 returned error can't find the container with id 5c8f20307c46ae6432f5adebbfa31b9b80e07a56e43541b298ce066b4f93197a Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.856414 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 15:56:59 crc kubenswrapper[4880]: I0126 15:56:59.865208 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.184340 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:00 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:00 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:00 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.184796 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.202586 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 15:57:00 crc kubenswrapper[4880]: W0126 15:57:00.214372 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e2725c4_874b_4e90_a0c1_ade113a0b7e2.slice/crio-89e2a3313aa26f2668d7aef6cb58afbb222f543405db70d5b8e27ae349457469 WatchSource:0}: Error finding container 89e2a3313aa26f2668d7aef6cb58afbb222f543405db70d5b8e27ae349457469: Status 404 returned error can't find the container with id 89e2a3313aa26f2668d7aef6cb58afbb222f543405db70d5b8e27ae349457469 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.221668 4880 generic.go:334] "Generic (PLEG): container finished" podID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerID="a2cfea7242b4033b4ee6d62088217819b86613d2c676ff386f0d1e23f0b37dec" exitCode=0 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.229150 4880 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.230290 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.231331 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerDied","Data":"a2cfea7242b4033b4ee6d62088217819b86613d2c676ff386f0d1e23f0b37dec"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.231398 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerStarted","Data":"5c8f20307c46ae6432f5adebbfa31b9b80e07a56e43541b298ce066b4f93197a"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.234954 4880 generic.go:334] "Generic (PLEG): container finished" podID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerID="3255312287a4bb286b16d259765e4172f95bdf38861a58674ade6c60485e494b" exitCode=0 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.235030 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerDied","Data":"3255312287a4bb286b16d259765e4172f95bdf38861a58674ade6c60485e494b"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.235096 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerStarted","Data":"5be71e230c6932e24882571670740b203d4ca63cb3f0c58646a59c4cda9ea17c"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.244113 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" event={"ID":"c631c684-80bc-4f57-ae38-bb96c0db15d1","Type":"ContainerStarted","Data":"0ffc6ed32a0af8a9cc7eb9ef3633b6710bf1fd5d0ad2081864a482e707f764dc"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.244171 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" event={"ID":"c631c684-80bc-4f57-ae38-bb96c0db15d1","Type":"ContainerStarted","Data":"1382b7ddf46ac986b7c99562adb436f93d0c757c4a1f3768d248db8968f296fc"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.269491 4880 generic.go:334] "Generic (PLEG): container finished" podID="35862492-137b-47ad-9547-9e54cb19674c" containerID="e802536ad089063a58c1ec07c30c43a8d11c20f2b172048560c4d2c10849d7f4" exitCode=0 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.269604 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"35862492-137b-47ad-9547-9e54cb19674c","Type":"ContainerDied","Data":"e802536ad089063a58c1ec07c30c43a8d11c20f2b172048560c4d2c10849d7f4"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.276989 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pxxnr" podStartSLOduration=31.276958394 podStartE2EDuration="31.276958394s" podCreationTimestamp="2026-01-26 15:56:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:57:00.274509334 +0000 UTC m=+145.760238041" watchObservedRunningTime="2026-01-26 15:57:00.276958394 +0000 UTC m=+145.762687101" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.278114 4880 generic.go:334] "Generic (PLEG): container finished" podID="7019dea6-6033-4097-9410-092fe22a8402" containerID="250d2d8746f26b20afce5879f982c505109cc6c87c62b1703ca170bb8e5c7bcc" exitCode=0 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.279150 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerDied","Data":"250d2d8746f26b20afce5879f982c505109cc6c87c62b1703ca170bb8e5c7bcc"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.279239 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerStarted","Data":"baf400fe29a08710b74f11311d99f507d2d82e59354895c536785430dd8cf41e"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.288452 4880 generic.go:334] "Generic (PLEG): container finished" podID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerID="3f580137fad1bdca344dcf575ca400505815ad0151674bcdac16af4f8daac7c3" exitCode=0 Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.289686 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerDied","Data":"3f580137fad1bdca344dcf575ca400505815ad0151674bcdac16af4f8daac7c3"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.289717 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerStarted","Data":"80e0f9ba0d31d0cbc5f110a12a47bcf7ea1ca2f381a15d5d2b5673884e0fa6bf"} Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.839868 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.843035 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.850642 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.893456 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.893531 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhhm7\" (UniqueName: \"kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.893595 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.899641 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.900310 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.995018 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.995395 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.995455 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhhm7\" (UniqueName: \"kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.996327 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:00 crc kubenswrapper[4880]: I0126 15:57:00.996355 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.019502 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhhm7\" (UniqueName: \"kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7\") pod \"redhat-marketplace-r4dp4\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.028947 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cg95b" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.096550 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume\") pod \"de011de4-f35b-4f67-b436-c965f427dbf2\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.096687 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume\") pod \"de011de4-f35b-4f67-b436-c965f427dbf2\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.096757 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzgv7\" (UniqueName: \"kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7\") pod \"de011de4-f35b-4f67-b436-c965f427dbf2\" (UID: \"de011de4-f35b-4f67-b436-c965f427dbf2\") " Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.097993 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume" (OuterVolumeSpecName: "config-volume") pod "de011de4-f35b-4f67-b436-c965f427dbf2" (UID: "de011de4-f35b-4f67-b436-c965f427dbf2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.101335 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7" (OuterVolumeSpecName: "kube-api-access-bzgv7") pod "de011de4-f35b-4f67-b436-c965f427dbf2" (UID: "de011de4-f35b-4f67-b436-c965f427dbf2"). InnerVolumeSpecName "kube-api-access-bzgv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.101372 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "de011de4-f35b-4f67-b436-c965f427dbf2" (UID: "de011de4-f35b-4f67-b436-c965f427dbf2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.183021 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:01 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:01 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:01 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.183111 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.199195 4880 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de011de4-f35b-4f67-b436-c965f427dbf2-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.199270 4880 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de011de4-f35b-4f67-b436-c965f427dbf2-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.200142 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzgv7\" (UniqueName: \"kubernetes.io/projected/de011de4-f35b-4f67-b436-c965f427dbf2-kube-api-access-bzgv7\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.221946 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:57:01 crc kubenswrapper[4880]: E0126 15:57:01.222577 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de011de4-f35b-4f67-b436-c965f427dbf2" containerName="collect-profiles" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.222773 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="de011de4-f35b-4f67-b436-c965f427dbf2" containerName="collect-profiles" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.222990 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="de011de4-f35b-4f67-b436-c965f427dbf2" containerName="collect-profiles" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.223909 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.227756 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.236065 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.290134 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.309517 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" event={"ID":"5e2725c4-874b-4e90-a0c1-ade113a0b7e2","Type":"ContainerStarted","Data":"89e2a3313aa26f2668d7aef6cb58afbb222f543405db70d5b8e27ae349457469"} Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.315701 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" event={"ID":"de011de4-f35b-4f67-b436-c965f427dbf2","Type":"ContainerDied","Data":"fcc2519ad75d0cc1a211b6f8e81220dd79383c619a8fbdf40c464da938e53ea9"} Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.315831 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcc2519ad75d0cc1a211b6f8e81220dd79383c619a8fbdf40c464da938e53ea9" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.316019 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490705-qzt6g" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.417224 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.417724 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.417929 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qlvw\" (UniqueName: \"kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.519697 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.519792 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.519846 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qlvw\" (UniqueName: \"kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.520845 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.521148 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.543077 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.561770 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qlvw\" (UniqueName: \"kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw\") pod \"redhat-marketplace-sh2cm\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.631411 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.636240 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.638124 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.649959 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.667047 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.728241 4880 patch_prober.go:28] interesting pod/console-f9d7485db-6ft2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.728344 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6ft2z" podUID="c8272257-ba55-49eb-84c5-65cf2987c8ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824214 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir\") pod \"35862492-137b-47ad-9547-9e54cb19674c\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824293 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access\") pod \"35862492-137b-47ad-9547-9e54cb19674c\" (UID: \"35862492-137b-47ad-9547-9e54cb19674c\") " Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824523 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqd95\" (UniqueName: \"kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824609 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824672 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.824777 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "35862492-137b-47ad-9547-9e54cb19674c" (UID: "35862492-137b-47ad-9547-9e54cb19674c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.826859 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:57:01 crc kubenswrapper[4880]: E0126 15:57:01.827206 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35862492-137b-47ad-9547-9e54cb19674c" containerName="pruner" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.827234 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="35862492-137b-47ad-9547-9e54cb19674c" containerName="pruner" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.827378 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="35862492-137b-47ad-9547-9e54cb19674c" containerName="pruner" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.828532 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.831087 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "35862492-137b-47ad-9547-9e54cb19674c" (UID: "35862492-137b-47ad-9547-9e54cb19674c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.853582 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.853946 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.925803 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.926008 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.926751 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.926787 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.927597 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqd95\" (UniqueName: \"kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.927738 4880 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35862492-137b-47ad-9547-9e54cb19674c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.927762 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35862492-137b-47ad-9547-9e54cb19674c-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.967754 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqd95\" (UniqueName: \"kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95\") pod \"redhat-operators-qr9hz\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:01 crc kubenswrapper[4880]: I0126 15:57:01.996298 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.029493 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zmtj\" (UniqueName: \"kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.029645 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.029901 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.133227 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zmtj\" (UniqueName: \"kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.133318 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.133341 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.134098 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.134809 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.174345 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zmtj\" (UniqueName: \"kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj\") pod \"redhat-operators-wp6pg\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.186834 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:02 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:02 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:02 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.186918 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.192513 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.192586 4880 patch_prober.go:28] interesting pod/downloads-7954f5f757-2wj2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.192609 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.192655 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-2wj2q" podUID="c0fb73bb-2bd2-460f-9d46-a1b44680e187" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.192803 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4vdpp" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.199349 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.247185 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-h6x5b" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.258365 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.260916 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.269755 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.274321 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.306208 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.342533 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" event={"ID":"5e2725c4-874b-4e90-a0c1-ade113a0b7e2","Type":"ContainerStarted","Data":"38eca27fc527afe3bb1284d98916ecd88c77aa5969c29ead80b761ff6ee37bb9"} Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.342963 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.350083 4880 generic.go:334] "Generic (PLEG): container finished" podID="60d83505-44fa-47d9-b05c-def5eda227a9" containerID="599a6329ec7869ff418b2aae242e419ecde6e7a5313fe73db51f41e7943cdd49" exitCode=0 Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.350475 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerDied","Data":"599a6329ec7869ff418b2aae242e419ecde6e7a5313fe73db51f41e7943cdd49"} Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.350680 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerStarted","Data":"80554d42406f1e7cd460e45eaa71a861c4c8b562723a42c8dc451b8d3ba447cb"} Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.363655 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.363655 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"35862492-137b-47ad-9547-9e54cb19674c","Type":"ContainerDied","Data":"ec5a093822e4bb60b6c4c8ffd6869fd79f9bd43992893c9b8dbef04f510c765c"} Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.364533 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec5a093822e4bb60b6c4c8ffd6869fd79f9bd43992893c9b8dbef04f510c765c" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.437046 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" podStartSLOduration=113.437018325 podStartE2EDuration="1m53.437018325s" podCreationTimestamp="2026-01-26 15:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:57:02.427340948 +0000 UTC m=+147.913069655" watchObservedRunningTime="2026-01-26 15:57:02.437018325 +0000 UTC m=+147.922747032" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.456399 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.456553 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.458634 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.500508 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.557818 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.558815 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.559156 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.599004 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.609116 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.682899 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 15:57:02 crc kubenswrapper[4880]: W0126 15:57:02.748556 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b678a2b_2601_4267_8f78_9b6262b8261a.slice/crio-f06cd8f4ea781e781cab769961c1b1286a9fddd367cc2fd2791990d0ae297321 WatchSource:0}: Error finding container f06cd8f4ea781e781cab769961c1b1286a9fddd367cc2fd2791990d0ae297321: Status 404 returned error can't find the container with id f06cd8f4ea781e781cab769961c1b1286a9fddd367cc2fd2791990d0ae297321 Jan 26 15:57:02 crc kubenswrapper[4880]: I0126 15:57:02.978007 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:57:03 crc kubenswrapper[4880]: W0126 15:57:03.025752 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61a4c42a_99ee_453e_b551_e29b8893a44d.slice/crio-9c2c209cfb28dbae77d9fcbaef9f75c0f4d6eded35742055d65736a57f517a58 WatchSource:0}: Error finding container 9c2c209cfb28dbae77d9fcbaef9f75c0f4d6eded35742055d65736a57f517a58: Status 404 returned error can't find the container with id 9c2c209cfb28dbae77d9fcbaef9f75c0f4d6eded35742055d65736a57f517a58 Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.191325 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:03 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:03 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:03 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.191450 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.285855 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 15:57:03 crc kubenswrapper[4880]: W0126 15:57:03.340973 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod35f869b8_3461_44f0_92e4_cd0f797081fd.slice/crio-2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5 WatchSource:0}: Error finding container 2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5: Status 404 returned error can't find the container with id 2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5 Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.378333 4880 generic.go:334] "Generic (PLEG): container finished" podID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerID="08daebfe2d6811c42521510ead7d6c2d2d043b01ff1bdd8de86c5a6dd707dc76" exitCode=0 Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.378382 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerDied","Data":"08daebfe2d6811c42521510ead7d6c2d2d043b01ff1bdd8de86c5a6dd707dc76"} Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.378446 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerStarted","Data":"f06cd8f4ea781e781cab769961c1b1286a9fddd367cc2fd2791990d0ae297321"} Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.380952 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"35f869b8-3461-44f0-92e4-cd0f797081fd","Type":"ContainerStarted","Data":"2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5"} Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.384406 4880 generic.go:334] "Generic (PLEG): container finished" podID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerID="3247f6267b406146e28601fa764386dacf11433fca8e05649c48c9f6b8eea710" exitCode=0 Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.384471 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerDied","Data":"3247f6267b406146e28601fa764386dacf11433fca8e05649c48c9f6b8eea710"} Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.384508 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerStarted","Data":"c9d93b585d5db9987567f50e0488aacc93974bb26c1a22fab59ea2895f93ed02"} Jan 26 15:57:03 crc kubenswrapper[4880]: I0126 15:57:03.386323 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerStarted","Data":"9c2c209cfb28dbae77d9fcbaef9f75c0f4d6eded35742055d65736a57f517a58"} Jan 26 15:57:04 crc kubenswrapper[4880]: I0126 15:57:04.305369 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:04 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:04 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:04 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:04 crc kubenswrapper[4880]: I0126 15:57:04.306555 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:04 crc kubenswrapper[4880]: I0126 15:57:04.572087 4880 generic.go:334] "Generic (PLEG): container finished" podID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerID="1b843f2a0849371dd3fd28e2de357cf7f8743d82c68d653eaf7a4c8108c08509" exitCode=0 Jan 26 15:57:04 crc kubenswrapper[4880]: I0126 15:57:04.572168 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerDied","Data":"1b843f2a0849371dd3fd28e2de357cf7f8743d82c68d653eaf7a4c8108c08509"} Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.182455 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:05 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:05 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:05 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.183058 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.258020 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.626914 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"35f869b8-3461-44f0-92e4-cd0f797081fd","Type":"ContainerStarted","Data":"c71b58e1a53b8b888929d9fef9c45fda6d15d8499e9b4bb1f9c53c311b9cb706"} Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.650923 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.650885018 podStartE2EDuration="3.650885018s" podCreationTimestamp="2026-01-26 15:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:57:05.646095101 +0000 UTC m=+151.131823818" watchObservedRunningTime="2026-01-26 15:57:05.650885018 +0000 UTC m=+151.136613725" Jan 26 15:57:05 crc kubenswrapper[4880]: I0126 15:57:05.780184 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-gmdcv" Jan 26 15:57:06 crc kubenswrapper[4880]: I0126 15:57:06.210896 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:06 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:06 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:06 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:06 crc kubenswrapper[4880]: I0126 15:57:06.211056 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:06 crc kubenswrapper[4880]: I0126 15:57:06.645149 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"35f869b8-3461-44f0-92e4-cd0f797081fd","Type":"ContainerDied","Data":"c71b58e1a53b8b888929d9fef9c45fda6d15d8499e9b4bb1f9c53c311b9cb706"} Jan 26 15:57:06 crc kubenswrapper[4880]: I0126 15:57:06.645147 4880 generic.go:334] "Generic (PLEG): container finished" podID="35f869b8-3461-44f0-92e4-cd0f797081fd" containerID="c71b58e1a53b8b888929d9fef9c45fda6d15d8499e9b4bb1f9c53c311b9cb706" exitCode=0 Jan 26 15:57:07 crc kubenswrapper[4880]: I0126 15:57:07.181977 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:07 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:07 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:07 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:07 crc kubenswrapper[4880]: I0126 15:57:07.182164 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.131898 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.161585 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access\") pod \"35f869b8-3461-44f0-92e4-cd0f797081fd\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.161737 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir\") pod \"35f869b8-3461-44f0-92e4-cd0f797081fd\" (UID: \"35f869b8-3461-44f0-92e4-cd0f797081fd\") " Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.162490 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "35f869b8-3461-44f0-92e4-cd0f797081fd" (UID: "35f869b8-3461-44f0-92e4-cd0f797081fd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.184654 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:08 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:08 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:08 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.184795 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.200574 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "35f869b8-3461-44f0-92e4-cd0f797081fd" (UID: "35f869b8-3461-44f0-92e4-cd0f797081fd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.267367 4880 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/35f869b8-3461-44f0-92e4-cd0f797081fd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:08 crc kubenswrapper[4880]: I0126 15:57:08.267410 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/35f869b8-3461-44f0-92e4-cd0f797081fd-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:57:09 crc kubenswrapper[4880]: I0126 15:57:09.024757 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"35f869b8-3461-44f0-92e4-cd0f797081fd","Type":"ContainerDied","Data":"2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5"} Jan 26 15:57:09 crc kubenswrapper[4880]: I0126 15:57:09.031616 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b44b5111af5a7aac886277d426b520d223369d2e2386ca52c21712046385de5" Jan 26 15:57:09 crc kubenswrapper[4880]: I0126 15:57:09.031765 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 15:57:09 crc kubenswrapper[4880]: I0126 15:57:09.181719 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:09 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:09 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:09 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:09 crc kubenswrapper[4880]: I0126 15:57:09.181827 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:10 crc kubenswrapper[4880]: I0126 15:57:10.182668 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:10 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:10 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:10 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:10 crc kubenswrapper[4880]: I0126 15:57:10.183123 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:10 crc kubenswrapper[4880]: I0126 15:57:10.701477 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 15:57:10 crc kubenswrapper[4880]: I0126 15:57:10.701569 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 15:57:11 crc kubenswrapper[4880]: I0126 15:57:11.354871 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:11 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:11 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:11 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:11 crc kubenswrapper[4880]: I0126 15:57:11.355267 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:11 crc kubenswrapper[4880]: I0126 15:57:11.732719 4880 patch_prober.go:28] interesting pod/console-f9d7485db-6ft2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 26 15:57:11 crc kubenswrapper[4880]: I0126 15:57:11.732805 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6ft2z" podUID="c8272257-ba55-49eb-84c5-65cf2987c8ce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 26 15:57:12 crc kubenswrapper[4880]: I0126 15:57:12.182712 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:12 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:12 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:12 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:12 crc kubenswrapper[4880]: I0126 15:57:12.182783 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:12 crc kubenswrapper[4880]: I0126 15:57:12.242292 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-2wj2q" Jan 26 15:57:12 crc kubenswrapper[4880]: I0126 15:57:12.242391 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hvhpp" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.453553 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.453975 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.454030 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.461718 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:13 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:13 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:13 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.464659 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.463248 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.463633 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.463685 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.474357 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.476535 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.479491 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.517893 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.839778 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:57:13 crc kubenswrapper[4880]: I0126 15:57:13.840232 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 15:57:14 crc kubenswrapper[4880]: I0126 15:57:14.332326 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:57:14 crc kubenswrapper[4880]: I0126 15:57:14.336849 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:57:14 crc kubenswrapper[4880]: I0126 15:57:14.339846 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 15:57:14 crc kubenswrapper[4880]: I0126 15:57:14.350490 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:14 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:14 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:14 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:14 crc kubenswrapper[4880]: I0126 15:57:14.350580 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:15 crc kubenswrapper[4880]: I0126 15:57:15.181942 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:15 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:15 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:15 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:15 crc kubenswrapper[4880]: I0126 15:57:15.182045 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:16 crc kubenswrapper[4880]: I0126 15:57:16.181477 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:16 crc kubenswrapper[4880]: [-]has-synced failed: reason withheld Jan 26 15:57:16 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:16 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:16 crc kubenswrapper[4880]: I0126 15:57:16.181575 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:17 crc kubenswrapper[4880]: I0126 15:57:17.183098 4880 patch_prober.go:28] interesting pod/router-default-5444994796-d4nqf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 15:57:17 crc kubenswrapper[4880]: [+]has-synced ok Jan 26 15:57:17 crc kubenswrapper[4880]: [+]process-running ok Jan 26 15:57:17 crc kubenswrapper[4880]: healthz check failed Jan 26 15:57:17 crc kubenswrapper[4880]: I0126 15:57:17.183182 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d4nqf" podUID="9340b971-8db7-4e70-bbc9-0e4d7292b4e1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:57:18 crc kubenswrapper[4880]: I0126 15:57:18.182850 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:57:18 crc kubenswrapper[4880]: I0126 15:57:18.186808 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-d4nqf" Jan 26 15:57:19 crc kubenswrapper[4880]: I0126 15:57:19.871493 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 15:57:22 crc kubenswrapper[4880]: I0126 15:57:22.078318 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:57:22 crc kubenswrapper[4880]: I0126 15:57:22.083684 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6ft2z" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.634082 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 15:57:37 crc kubenswrapper[4880]: E0126 15:57:37.635260 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f869b8-3461-44f0-92e4-cd0f797081fd" containerName="pruner" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.635288 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f869b8-3461-44f0-92e4-cd0f797081fd" containerName="pruner" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.635520 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f869b8-3461-44f0-92e4-cd0f797081fd" containerName="pruner" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.636512 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.640680 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.640728 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 15:57:37 crc kubenswrapper[4880]: I0126 15:57:37.644075 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.023429 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.023579 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.125268 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.125425 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.125902 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.145546 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:38 crc kubenswrapper[4880]: I0126 15:57:38.248059 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:57:40 crc kubenswrapper[4880]: I0126 15:57:40.700162 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 15:57:40 crc kubenswrapper[4880]: I0126 15:57:40.700571 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.020999 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.022231 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.039367 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.067538 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.067658 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.067715 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.169828 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.170154 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.170226 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.170317 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.170494 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.191025 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access\") pod \"installer-9-crc\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:43 crc kubenswrapper[4880]: I0126 15:57:43.346572 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:57:52 crc kubenswrapper[4880]: E0126 15:57:52.889007 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 15:57:52 crc kubenswrapper[4880]: E0126 15:57:52.889871 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wqfwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5d9pw_openshift-marketplace(7fff77b4-8896-4c5f-8960-93aedba5de59): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:52 crc kubenswrapper[4880]: E0126 15:57:52.891147 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5d9pw" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" Jan 26 15:57:54 crc kubenswrapper[4880]: E0126 15:57:54.132076 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5d9pw" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" Jan 26 15:57:54 crc kubenswrapper[4880]: E0126 15:57:54.204191 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 15:57:54 crc kubenswrapper[4880]: E0126 15:57:54.204597 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhhm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-r4dp4_openshift-marketplace(60d83505-44fa-47d9-b05c-def5eda227a9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:54 crc kubenswrapper[4880]: E0126 15:57:54.205938 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-r4dp4" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" Jan 26 15:57:58 crc kubenswrapper[4880]: E0126 15:57:58.069222 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-r4dp4" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" Jan 26 15:57:58 crc kubenswrapper[4880]: E0126 15:57:58.159611 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 15:57:58 crc kubenswrapper[4880]: E0126 15:57:58.159785 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sqd95,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qr9hz_openshift-marketplace(4b678a2b-2601-4267-8f78-9b6262b8261a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:58 crc kubenswrapper[4880]: E0126 15:57:58.161033 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qr9hz" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.617379 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qr9hz" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.688396 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.688753 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cjdcs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t9qds_openshift-marketplace(0f25591a-acee-4cc9-98b5-ca45d47ef1bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.690034 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t9qds" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.777227 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.777726 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4zmtj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-wp6pg_openshift-marketplace(61a4c42a-99ee-453e-b551-e29b8893a44d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.778938 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-wp6pg" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.780521 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.780652 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dvxx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r6zxd_openshift-marketplace(7019dea6-6033-4097-9410-092fe22a8402): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.781824 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r6zxd" podUID="7019dea6-6033-4097-9410-092fe22a8402" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.802406 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.802628 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bmdfr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4xxrk_openshift-marketplace(ebb3fdcf-80cb-4293-9e7d-17a44072c91e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.803818 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.804075 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8qlvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-sh2cm_openshift-marketplace(22218cc4-cad8-4764-ac76-d482abd7f2d6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.804257 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4xxrk" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" Jan 26 15:57:59 crc kubenswrapper[4880]: E0126 15:57:59.805485 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-sh2cm" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" Jan 26 15:58:00 crc kubenswrapper[4880]: E0126 15:58:00.126500 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4xxrk" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" Jan 26 15:58:00 crc kubenswrapper[4880]: E0126 15:58:00.126882 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r6zxd" podUID="7019dea6-6033-4097-9410-092fe22a8402" Jan 26 15:58:00 crc kubenswrapper[4880]: E0126 15:58:00.126960 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-wp6pg" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" Jan 26 15:58:00 crc kubenswrapper[4880]: E0126 15:58:00.128301 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t9qds" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" Jan 26 15:58:00 crc kubenswrapper[4880]: E0126 15:58:00.128465 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-sh2cm" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" Jan 26 15:58:00 crc kubenswrapper[4880]: W0126 15:58:00.170814 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-33bad18bd0119c5cccc459236b7607ae3ca52b7edd1d59262aff32eb76109514 WatchSource:0}: Error finding container 33bad18bd0119c5cccc459236b7607ae3ca52b7edd1d59262aff32eb76109514: Status 404 returned error can't find the container with id 33bad18bd0119c5cccc459236b7607ae3ca52b7edd1d59262aff32eb76109514 Jan 26 15:58:00 crc kubenswrapper[4880]: W0126 15:58:00.174143 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-29e981ef2522257ada987c5fc2c9ddb4b0c0bc1099886d8c57207a9880e72ccb WatchSource:0}: Error finding container 29e981ef2522257ada987c5fc2c9ddb4b0c0bc1099886d8c57207a9880e72ccb: Status 404 returned error can't find the container with id 29e981ef2522257ada987c5fc2c9ddb4b0c0bc1099886d8c57207a9880e72ccb Jan 26 15:58:00 crc kubenswrapper[4880]: I0126 15:58:00.174974 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 15:58:00 crc kubenswrapper[4880]: I0126 15:58:00.428990 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.133867 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"70af447ebe586257b4777ded54bceb18080fbdf71583a32ec29896daa01ca454"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.134294 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"33bad18bd0119c5cccc459236b7607ae3ca52b7edd1d59262aff32eb76109514"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.138062 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6802daa6-8d51-4a30-b60a-7583a3caa317","Type":"ContainerStarted","Data":"b48300f75aa39fb17689f7d78c521d23991e0c746ced0d1277eb2da6d85a8cb1"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.138134 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6802daa6-8d51-4a30-b60a-7583a3caa317","Type":"ContainerStarted","Data":"abefece38234e90bdace0403dcc93dffb12f83582cc1fa00b3a5f2a2728493e0"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.140167 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"09fb94098fb3de51ca2ed8a08ea475318625aaea516edb4dbb9917fb3c6dfa75"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.140268 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5c2e3b5aec4593c75c61a3387487588e62c46e480a444058debca34482a60ca7"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.141784 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7cc03a3909c7234e2d06a03d66c192ea9366ab539a2461430f50f40295d34795"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.141850 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"29e981ef2522257ada987c5fc2c9ddb4b0c0bc1099886d8c57207a9880e72ccb"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.142404 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.144191 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3f83e637-92b1-41a8-9df9-f1d5420ca1f8","Type":"ContainerStarted","Data":"63d2b6b8ae3778f6df81ac62ba335b9fe541abff855ebf132a9ae7b652d9ea61"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.144232 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3f83e637-92b1-41a8-9df9-f1d5420ca1f8","Type":"ContainerStarted","Data":"3431f25d8b899d8c92b1c7350a56f3b25c0bac0d43a45e04e7325275296db547"} Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.185865 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=24.185824593 podStartE2EDuration="24.185824593s" podCreationTimestamp="2026-01-26 15:57:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:58:01.175323459 +0000 UTC m=+206.661052186" watchObservedRunningTime="2026-01-26 15:58:01.185824593 +0000 UTC m=+206.671553300" Jan 26 15:58:01 crc kubenswrapper[4880]: I0126 15:58:01.220942 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=18.220920328 podStartE2EDuration="18.220920328s" podCreationTimestamp="2026-01-26 15:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:58:01.218082004 +0000 UTC m=+206.703810711" watchObservedRunningTime="2026-01-26 15:58:01.220920328 +0000 UTC m=+206.706649035" Jan 26 15:58:02 crc kubenswrapper[4880]: I0126 15:58:02.152060 4880 generic.go:334] "Generic (PLEG): container finished" podID="6802daa6-8d51-4a30-b60a-7583a3caa317" containerID="b48300f75aa39fb17689f7d78c521d23991e0c746ced0d1277eb2da6d85a8cb1" exitCode=0 Jan 26 15:58:02 crc kubenswrapper[4880]: I0126 15:58:02.152124 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6802daa6-8d51-4a30-b60a-7583a3caa317","Type":"ContainerDied","Data":"b48300f75aa39fb17689f7d78c521d23991e0c746ced0d1277eb2da6d85a8cb1"} Jan 26 15:58:02 crc kubenswrapper[4880]: I0126 15:58:02.744020 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.521617 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.631572 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access\") pod \"6802daa6-8d51-4a30-b60a-7583a3caa317\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.631707 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir\") pod \"6802daa6-8d51-4a30-b60a-7583a3caa317\" (UID: \"6802daa6-8d51-4a30-b60a-7583a3caa317\") " Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.631987 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6802daa6-8d51-4a30-b60a-7583a3caa317" (UID: "6802daa6-8d51-4a30-b60a-7583a3caa317"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.640652 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6802daa6-8d51-4a30-b60a-7583a3caa317" (UID: "6802daa6-8d51-4a30-b60a-7583a3caa317"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.733560 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6802daa6-8d51-4a30-b60a-7583a3caa317-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:03 crc kubenswrapper[4880]: I0126 15:58:03.733611 4880 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6802daa6-8d51-4a30-b60a-7583a3caa317-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:04 crc kubenswrapper[4880]: I0126 15:58:04.174311 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6802daa6-8d51-4a30-b60a-7583a3caa317","Type":"ContainerDied","Data":"abefece38234e90bdace0403dcc93dffb12f83582cc1fa00b3a5f2a2728493e0"} Jan 26 15:58:04 crc kubenswrapper[4880]: I0126 15:58:04.174698 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abefece38234e90bdace0403dcc93dffb12f83582cc1fa00b3a5f2a2728493e0" Jan 26 15:58:04 crc kubenswrapper[4880]: I0126 15:58:04.174484 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 15:58:10 crc kubenswrapper[4880]: I0126 15:58:10.703928 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 15:58:10 crc kubenswrapper[4880]: I0126 15:58:10.704035 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 15:58:10 crc kubenswrapper[4880]: I0126 15:58:10.704123 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 15:58:10 crc kubenswrapper[4880]: I0126 15:58:10.704928 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 15:58:10 crc kubenswrapper[4880]: I0126 15:58:10.705140 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed" gracePeriod=600 Jan 26 15:58:11 crc kubenswrapper[4880]: I0126 15:58:11.341856 4880 generic.go:334] "Generic (PLEG): container finished" podID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerID="74e62edc0a5e5329a7f77b265b3b3d6b05fc13b62716d944eb47b99b77e97d6d" exitCode=0 Jan 26 15:58:11 crc kubenswrapper[4880]: I0126 15:58:11.342386 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerDied","Data":"74e62edc0a5e5329a7f77b265b3b3d6b05fc13b62716d944eb47b99b77e97d6d"} Jan 26 15:58:11 crc kubenswrapper[4880]: I0126 15:58:11.347595 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed" exitCode=0 Jan 26 15:58:11 crc kubenswrapper[4880]: I0126 15:58:11.347698 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed"} Jan 26 15:58:11 crc kubenswrapper[4880]: I0126 15:58:11.347796 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a"} Jan 26 15:58:13 crc kubenswrapper[4880]: I0126 15:58:13.361500 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerStarted","Data":"894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0"} Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.369135 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerStarted","Data":"1a9b160c127c09e079cf22089510965ca10d002fe57638e426a5e663cbf684dd"} Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.370705 4880 generic.go:334] "Generic (PLEG): container finished" podID="60d83505-44fa-47d9-b05c-def5eda227a9" containerID="25e8984948d1d66f4e2dc7f79754d5b3cda3483855571b1a90af19a554227884" exitCode=0 Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.370759 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerDied","Data":"25e8984948d1d66f4e2dc7f79754d5b3cda3483855571b1a90af19a554227884"} Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.375936 4880 generic.go:334] "Generic (PLEG): container finished" podID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerID="10dc118dcd339d3a303a9a6aadb92b4049b7d0dd1c3b9c310503f32b6d925832" exitCode=0 Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.376057 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerDied","Data":"10dc118dcd339d3a303a9a6aadb92b4049b7d0dd1c3b9c310503f32b6d925832"} Jan 26 15:58:14 crc kubenswrapper[4880]: I0126 15:58:14.463600 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5d9pw" podStartSLOduration=3.580297535 podStartE2EDuration="1m16.46355854s" podCreationTimestamp="2026-01-26 15:56:58 +0000 UTC" firstStartedPulling="2026-01-26 15:57:00.238561844 +0000 UTC m=+145.724290551" lastFinishedPulling="2026-01-26 15:58:13.121822849 +0000 UTC m=+218.607551556" observedRunningTime="2026-01-26 15:58:14.459941225 +0000 UTC m=+219.945669922" watchObservedRunningTime="2026-01-26 15:58:14.46355854 +0000 UTC m=+219.949287247" Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.386697 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerStarted","Data":"e452e81b7bbf67ac671b2efa53989cd97a9498c3ffe7ac495aa8c93082ccc873"} Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.392324 4880 generic.go:334] "Generic (PLEG): container finished" podID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerID="95f8ea210ee99020c0c6911532f871c1875e7a2b4f705fddd0c034a7477eb92f" exitCode=0 Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.392386 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerDied","Data":"95f8ea210ee99020c0c6911532f871c1875e7a2b4f705fddd0c034a7477eb92f"} Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.396638 4880 generic.go:334] "Generic (PLEG): container finished" podID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerID="1a9b160c127c09e079cf22089510965ca10d002fe57638e426a5e663cbf684dd" exitCode=0 Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.396701 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerDied","Data":"1a9b160c127c09e079cf22089510965ca10d002fe57638e426a5e663cbf684dd"} Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.402043 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerStarted","Data":"2147c5413f25743ec2170f5006969eb3d9047ad21d57a3b4d7f87ee419bbf6a8"} Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.413504 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sh2cm" podStartSLOduration=3.103323515 podStartE2EDuration="1m14.413474792s" podCreationTimestamp="2026-01-26 15:57:01 +0000 UTC" firstStartedPulling="2026-01-26 15:57:03.575399728 +0000 UTC m=+149.061128445" lastFinishedPulling="2026-01-26 15:58:14.885551005 +0000 UTC m=+220.371279722" observedRunningTime="2026-01-26 15:58:15.409880879 +0000 UTC m=+220.895609586" watchObservedRunningTime="2026-01-26 15:58:15.413474792 +0000 UTC m=+220.899203509" Jan 26 15:58:15 crc kubenswrapper[4880]: I0126 15:58:15.463814 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r4dp4" podStartSLOduration=3.032572412 podStartE2EDuration="1m15.463791644s" podCreationTimestamp="2026-01-26 15:57:00 +0000 UTC" firstStartedPulling="2026-01-26 15:57:02.351919922 +0000 UTC m=+147.837648629" lastFinishedPulling="2026-01-26 15:58:14.783139154 +0000 UTC m=+220.268867861" observedRunningTime="2026-01-26 15:58:15.458850206 +0000 UTC m=+220.944578913" watchObservedRunningTime="2026-01-26 15:58:15.463791644 +0000 UTC m=+220.949520361" Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.411711 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerStarted","Data":"ad7cff0409979eda239a6ebfe4c93d18ec5967d61b2cb0b8146a9596b95cb2ab"} Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.415507 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerStarted","Data":"e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149"} Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.417677 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerStarted","Data":"c8f097d359ac84d3120abc811fea63db1d54fb74e3b5a9e5154ab6c980ee8dfe"} Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.419832 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerStarted","Data":"73cd4f146134d887516a8bfdcfd1026653df0964a1333bf770acc797996be2e1"} Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.486858 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wp6pg" podStartSLOduration=4.000074031 podStartE2EDuration="1m15.486838094s" podCreationTimestamp="2026-01-26 15:57:01 +0000 UTC" firstStartedPulling="2026-01-26 15:57:04.593620322 +0000 UTC m=+150.079349019" lastFinishedPulling="2026-01-26 15:58:16.080384375 +0000 UTC m=+221.566113082" observedRunningTime="2026-01-26 15:58:16.481747011 +0000 UTC m=+221.967475738" watchObservedRunningTime="2026-01-26 15:58:16.486838094 +0000 UTC m=+221.972566801" Jan 26 15:58:16 crc kubenswrapper[4880]: I0126 15:58:16.546160 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4xxrk" podStartSLOduration=2.98089682 podStartE2EDuration="1m18.546135911s" podCreationTimestamp="2026-01-26 15:56:58 +0000 UTC" firstStartedPulling="2026-01-26 15:57:00.290274629 +0000 UTC m=+145.776003336" lastFinishedPulling="2026-01-26 15:58:15.85551372 +0000 UTC m=+221.341242427" observedRunningTime="2026-01-26 15:58:16.54377168 +0000 UTC m=+222.029500387" watchObservedRunningTime="2026-01-26 15:58:16.546135911 +0000 UTC m=+222.031864618" Jan 26 15:58:17 crc kubenswrapper[4880]: I0126 15:58:17.426609 4880 generic.go:334] "Generic (PLEG): container finished" podID="7019dea6-6033-4097-9410-092fe22a8402" containerID="ad7cff0409979eda239a6ebfe4c93d18ec5967d61b2cb0b8146a9596b95cb2ab" exitCode=0 Jan 26 15:58:17 crc kubenswrapper[4880]: I0126 15:58:17.426681 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerDied","Data":"ad7cff0409979eda239a6ebfe4c93d18ec5967d61b2cb0b8146a9596b95cb2ab"} Jan 26 15:58:17 crc kubenswrapper[4880]: I0126 15:58:17.430928 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerStarted","Data":"60a5d5e0206253a2fcec31676d6e1f69f841fb8a61805cf0d72bc0b0fabb667c"} Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.439137 4880 generic.go:334] "Generic (PLEG): container finished" podID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerID="73cd4f146134d887516a8bfdcfd1026653df0964a1333bf770acc797996be2e1" exitCode=0 Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.439255 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerDied","Data":"73cd4f146134d887516a8bfdcfd1026653df0964a1333bf770acc797996be2e1"} Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.441600 4880 generic.go:334] "Generic (PLEG): container finished" podID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerID="60a5d5e0206253a2fcec31676d6e1f69f841fb8a61805cf0d72bc0b0fabb667c" exitCode=0 Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.441636 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerDied","Data":"60a5d5e0206253a2fcec31676d6e1f69f841fb8a61805cf0d72bc0b0fabb667c"} Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.771743 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.771824 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.987622 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:58:18 crc kubenswrapper[4880]: I0126 15:58:18.987712 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:58:19 crc kubenswrapper[4880]: I0126 15:58:19.228626 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:58:19 crc kubenswrapper[4880]: I0126 15:58:19.230698 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:58:19 crc kubenswrapper[4880]: I0126 15:58:19.489579 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.228685 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.228999 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.277037 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.515168 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.855117 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.855183 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:21 crc kubenswrapper[4880]: I0126 15:58:21.906843 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:22 crc kubenswrapper[4880]: I0126 15:58:22.459388 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:22 crc kubenswrapper[4880]: I0126 15:58:22.459491 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:22 crc kubenswrapper[4880]: I0126 15:58:22.548617 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:22 crc kubenswrapper[4880]: I0126 15:58:22.555412 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:22 crc kubenswrapper[4880]: I0126 15:58:22.600960 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.462370 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.462979 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sh2cm" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="registry-server" containerID="cri-o://e452e81b7bbf67ac671b2efa53989cd97a9498c3ffe7ac495aa8c93082ccc873" gracePeriod=2 Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.544514 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerStarted","Data":"a02f9f7f40969bfc503b2362a64469cd69d300a0d9be53bbed04741e50a7a8d6"} Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.552498 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerStarted","Data":"30c7c6d17d43b2796d289522c26623a2bf6197f2d1b56093a53c0a73db9c52af"} Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.560321 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerStarted","Data":"8c21ea9d2eb01a3d364d798ebf077f5569b4e357b18e834c617f428cdc94185f"} Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.695405 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.695922 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wp6pg" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="registry-server" containerID="cri-o://c8f097d359ac84d3120abc811fea63db1d54fb74e3b5a9e5154ab6c980ee8dfe" gracePeriod=2 Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.698346 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t9qds" podStartSLOduration=1.8480651529999998 podStartE2EDuration="1m26.698303956s" podCreationTimestamp="2026-01-26 15:56:59 +0000 UTC" firstStartedPulling="2026-01-26 15:57:00.228741703 +0000 UTC m=+145.714470410" lastFinishedPulling="2026-01-26 15:58:25.078980506 +0000 UTC m=+230.564709213" observedRunningTime="2026-01-26 15:58:25.695942425 +0000 UTC m=+231.181671132" watchObservedRunningTime="2026-01-26 15:58:25.698303956 +0000 UTC m=+231.184032663" Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.883221 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r6zxd" podStartSLOduration=3.125374083 podStartE2EDuration="1m27.883204398s" podCreationTimestamp="2026-01-26 15:56:58 +0000 UTC" firstStartedPulling="2026-01-26 15:57:00.280677115 +0000 UTC m=+145.766405822" lastFinishedPulling="2026-01-26 15:58:25.03850743 +0000 UTC m=+230.524236137" observedRunningTime="2026-01-26 15:58:25.856756948 +0000 UTC m=+231.342485655" watchObservedRunningTime="2026-01-26 15:58:25.883204398 +0000 UTC m=+231.368933105" Jan 26 15:58:25 crc kubenswrapper[4880]: I0126 15:58:25.887018 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qr9hz" podStartSLOduration=3.368990944 podStartE2EDuration="1m24.886993667s" podCreationTimestamp="2026-01-26 15:57:01 +0000 UTC" firstStartedPulling="2026-01-26 15:57:03.5754726 +0000 UTC m=+149.061201307" lastFinishedPulling="2026-01-26 15:58:25.093475323 +0000 UTC m=+230.579204030" observedRunningTime="2026-01-26 15:58:25.882334645 +0000 UTC m=+231.368063352" watchObservedRunningTime="2026-01-26 15:58:25.886993667 +0000 UTC m=+231.372722364" Jan 26 15:58:26 crc kubenswrapper[4880]: I0126 15:58:26.600103 4880 generic.go:334] "Generic (PLEG): container finished" podID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerID="e452e81b7bbf67ac671b2efa53989cd97a9498c3ffe7ac495aa8c93082ccc873" exitCode=0 Jan 26 15:58:26 crc kubenswrapper[4880]: I0126 15:58:26.600206 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerDied","Data":"e452e81b7bbf67ac671b2efa53989cd97a9498c3ffe7ac495aa8c93082ccc873"} Jan 26 15:58:26 crc kubenswrapper[4880]: I0126 15:58:26.984696 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.114562 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content\") pod \"22218cc4-cad8-4764-ac76-d482abd7f2d6\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.216260 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qlvw\" (UniqueName: \"kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw\") pod \"22218cc4-cad8-4764-ac76-d482abd7f2d6\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.216388 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities\") pod \"22218cc4-cad8-4764-ac76-d482abd7f2d6\" (UID: \"22218cc4-cad8-4764-ac76-d482abd7f2d6\") " Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.217382 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities" (OuterVolumeSpecName: "utilities") pod "22218cc4-cad8-4764-ac76-d482abd7f2d6" (UID: "22218cc4-cad8-4764-ac76-d482abd7f2d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.225527 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw" (OuterVolumeSpecName: "kube-api-access-8qlvw") pod "22218cc4-cad8-4764-ac76-d482abd7f2d6" (UID: "22218cc4-cad8-4764-ac76-d482abd7f2d6"). InnerVolumeSpecName "kube-api-access-8qlvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.331264 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qlvw\" (UniqueName: \"kubernetes.io/projected/22218cc4-cad8-4764-ac76-d482abd7f2d6-kube-api-access-8qlvw\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.331350 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.608392 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sh2cm" event={"ID":"22218cc4-cad8-4764-ac76-d482abd7f2d6","Type":"ContainerDied","Data":"c9d93b585d5db9987567f50e0488aacc93974bb26c1a22fab59ea2895f93ed02"} Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.608518 4880 scope.go:117] "RemoveContainer" containerID="e452e81b7bbf67ac671b2efa53989cd97a9498c3ffe7ac495aa8c93082ccc873" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.608716 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sh2cm" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.619421 4880 generic.go:334] "Generic (PLEG): container finished" podID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerID="c8f097d359ac84d3120abc811fea63db1d54fb74e3b5a9e5154ab6c980ee8dfe" exitCode=0 Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.619515 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerDied","Data":"c8f097d359ac84d3120abc811fea63db1d54fb74e3b5a9e5154ab6c980ee8dfe"} Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.635236 4880 scope.go:117] "RemoveContainer" containerID="10dc118dcd339d3a303a9a6aadb92b4049b7d0dd1c3b9c310503f32b6d925832" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.659292 4880 scope.go:117] "RemoveContainer" containerID="3247f6267b406146e28601fa764386dacf11433fca8e05649c48c9f6b8eea710" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.723379 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22218cc4-cad8-4764-ac76-d482abd7f2d6" (UID: "22218cc4-cad8-4764-ac76-d482abd7f2d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.778275 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22218cc4-cad8-4764-ac76-d482abd7f2d6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:27 crc kubenswrapper[4880]: I0126 15:58:27.782877 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" containerID="cri-o://0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017" gracePeriod=15 Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.071072 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.082453 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sh2cm"] Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.221001 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.222764 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" path="/var/lib/kubelet/pods/22218cc4-cad8-4764-ac76-d482abd7f2d6/volumes" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.411087 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content\") pod \"61a4c42a-99ee-453e-b551-e29b8893a44d\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.411859 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities\") pod \"61a4c42a-99ee-453e-b551-e29b8893a44d\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.412004 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zmtj\" (UniqueName: \"kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj\") pod \"61a4c42a-99ee-453e-b551-e29b8893a44d\" (UID: \"61a4c42a-99ee-453e-b551-e29b8893a44d\") " Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.413231 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities" (OuterVolumeSpecName: "utilities") pod "61a4c42a-99ee-453e-b551-e29b8893a44d" (UID: "61a4c42a-99ee-453e-b551-e29b8893a44d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.425995 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj" (OuterVolumeSpecName: "kube-api-access-4zmtj") pod "61a4c42a-99ee-453e-b551-e29b8893a44d" (UID: "61a4c42a-99ee-453e-b551-e29b8893a44d"). InnerVolumeSpecName "kube-api-access-4zmtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.514118 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.514172 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zmtj\" (UniqueName: \"kubernetes.io/projected/61a4c42a-99ee-453e-b551-e29b8893a44d-kube-api-access-4zmtj\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.564138 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61a4c42a-99ee-453e-b551-e29b8893a44d" (UID: "61a4c42a-99ee-453e-b551-e29b8893a44d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.615486 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61a4c42a-99ee-453e-b551-e29b8893a44d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.627901 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp6pg" event={"ID":"61a4c42a-99ee-453e-b551-e29b8893a44d","Type":"ContainerDied","Data":"9c2c209cfb28dbae77d9fcbaef9f75c0f4d6eded35742055d65736a57f517a58"} Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.627961 4880 scope.go:117] "RemoveContainer" containerID="c8f097d359ac84d3120abc811fea63db1d54fb74e3b5a9e5154ab6c980ee8dfe" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.628066 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp6pg" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.645056 4880 scope.go:117] "RemoveContainer" containerID="1a9b160c127c09e079cf22089510965ca10d002fe57638e426a5e663cbf684dd" Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.657222 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.661097 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wp6pg"] Jan 26 15:58:28 crc kubenswrapper[4880]: I0126 15:58:28.685059 4880 scope.go:117] "RemoveContainer" containerID="1b843f2a0849371dd3fd28e2de357cf7f8743d82c68d653eaf7a4c8108c08509" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.029727 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.187969 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.188052 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.224708 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.521490 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.521533 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:58:29 crc kubenswrapper[4880]: I0126 15:58:29.566049 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.222956 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" path="/var/lib/kubelet/pods/61a4c42a-99ee-453e-b551-e29b8893a44d/volumes" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.232625 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.332828 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.332899 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.332943 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.332981 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333035 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333067 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333129 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333212 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333242 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333265 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333305 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333337 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333368 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncjt4\" (UniqueName: \"kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.333396 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca\") pod \"0d02c11b-b059-4d82-b864-5659fedb9fd5\" (UID: \"0d02c11b-b059-4d82-b864-5659fedb9fd5\") " Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.334547 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.337335 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.337702 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.338183 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.339641 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4" (OuterVolumeSpecName: "kube-api-access-ncjt4") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "kube-api-access-ncjt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.340876 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.341306 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.341872 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.342243 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.343973 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.344040 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.344390 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.344659 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.345583 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0d02c11b-b059-4d82-b864-5659fedb9fd5" (UID: "0d02c11b-b059-4d82-b864-5659fedb9fd5"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434592 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434664 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncjt4\" (UniqueName: \"kubernetes.io/projected/0d02c11b-b059-4d82-b864-5659fedb9fd5-kube-api-access-ncjt4\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434675 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434688 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434702 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434712 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434724 4880 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434735 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434748 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434759 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434769 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434779 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434789 4880 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.434816 4880 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d02c11b-b059-4d82-b864-5659fedb9fd5-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.651589 4880 generic.go:334] "Generic (PLEG): container finished" podID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerID="0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017" exitCode=0 Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.651723 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.651702 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" event={"ID":"0d02c11b-b059-4d82-b864-5659fedb9fd5","Type":"ContainerDied","Data":"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017"} Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.651881 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pxfwl" event={"ID":"0d02c11b-b059-4d82-b864-5659fedb9fd5","Type":"ContainerDied","Data":"3172857e23ed5b2476be953f0b998d8081605d7e9f69438cb8f96c1b307def40"} Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.651907 4880 scope.go:117] "RemoveContainer" containerID="0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.682871 4880 scope.go:117] "RemoveContainer" containerID="0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017" Jan 26 15:58:30 crc kubenswrapper[4880]: E0126 15:58:30.683551 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017\": container with ID starting with 0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017 not found: ID does not exist" containerID="0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.683588 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017"} err="failed to get container status \"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017\": rpc error: code = NotFound desc = could not find container \"0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017\": container with ID starting with 0bba101fd7dc103fb725a73687ba8e5186a1a02009c67769ee0ea4c20fa14017 not found: ID does not exist" Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.690597 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:58:30 crc kubenswrapper[4880]: I0126 15:58:30.693747 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pxfwl"] Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.380372 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-754dc54bdd-qk4x6"] Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.381771 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="extract-utilities" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.381922 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="extract-utilities" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382042 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="extract-content" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.382132 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="extract-content" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382217 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.382294 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382393 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="extract-content" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.382496 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="extract-content" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382621 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6802daa6-8d51-4a30-b60a-7583a3caa317" containerName="pruner" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.382731 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="6802daa6-8d51-4a30-b60a-7583a3caa317" containerName="pruner" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382812 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.382887 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.382951 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="extract-utilities" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383021 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="extract-utilities" Jan 26 15:58:31 crc kubenswrapper[4880]: E0126 15:58:31.383097 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383171 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383412 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="6802daa6-8d51-4a30-b60a-7583a3caa317" containerName="pruner" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383576 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="22218cc4-cad8-4764-ac76-d482abd7f2d6" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383646 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="61a4c42a-99ee-453e-b551-e29b8893a44d" containerName="registry-server" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.383732 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" containerName="oauth-openshift" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.384531 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.386904 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.387844 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.388679 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.389358 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.390283 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.390405 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.390778 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.391166 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.391543 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.391852 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.392247 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.392859 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.405947 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.408462 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.410994 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-754dc54bdd-qk4x6"] Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.418296 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563195 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-login\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563314 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-error\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563349 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-dir\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563378 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563424 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563485 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42fg\" (UniqueName: \"kubernetes.io/projected/77beb073-82ea-48a0-92c8-5d0d107638a8-kube-api-access-x42fg\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563510 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563549 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563596 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563644 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-policies\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563677 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-session\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563721 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563760 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.563790 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665001 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665067 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665120 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-policies\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665167 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-session\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665225 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665323 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665353 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665392 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-login\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665465 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-error\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665492 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-dir\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665537 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665590 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665617 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42fg\" (UniqueName: \"kubernetes.io/projected/77beb073-82ea-48a0-92c8-5d0d107638a8-kube-api-access-x42fg\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.665663 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.666821 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-dir\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.666937 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.667618 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.667713 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-audit-policies\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.670418 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.670842 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.671373 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.672275 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.677868 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.679226 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-error\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.679618 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-system-session\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.680164 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-login\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.683819 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/77beb073-82ea-48a0-92c8-5d0d107638a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.692876 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42fg\" (UniqueName: \"kubernetes.io/projected/77beb073-82ea-48a0-92c8-5d0d107638a8-kube-api-access-x42fg\") pod \"oauth-openshift-754dc54bdd-qk4x6\" (UID: \"77beb073-82ea-48a0-92c8-5d0d107638a8\") " pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.707696 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.997848 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:58:31 crc kubenswrapper[4880]: I0126 15:58:31.998161 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:58:32 crc kubenswrapper[4880]: I0126 15:58:32.148670 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-754dc54bdd-qk4x6"] Jan 26 15:58:32 crc kubenswrapper[4880]: I0126 15:58:32.222795 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d02c11b-b059-4d82-b864-5659fedb9fd5" path="/var/lib/kubelet/pods/0d02c11b-b059-4d82-b864-5659fedb9fd5/volumes" Jan 26 15:58:32 crc kubenswrapper[4880]: I0126 15:58:32.669557 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" event={"ID":"77beb073-82ea-48a0-92c8-5d0d107638a8","Type":"ContainerStarted","Data":"bc62133be352601a95619617ee50788c94ce14649049f4a1fb167db23ccfa69b"} Jan 26 15:58:33 crc kubenswrapper[4880]: I0126 15:58:33.039025 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qr9hz" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="registry-server" probeResult="failure" output=< Jan 26 15:58:33 crc kubenswrapper[4880]: timeout: failed to connect service ":50051" within 1s Jan 26 15:58:33 crc kubenswrapper[4880]: > Jan 26 15:58:35 crc kubenswrapper[4880]: I0126 15:58:35.537187 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 15:58:35 crc kubenswrapper[4880]: I0126 15:58:35.686963 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" event={"ID":"77beb073-82ea-48a0-92c8-5d0d107638a8","Type":"ContainerStarted","Data":"a988f418d5863577e493bcb2bf4ecee3bae19e125874e4d2a1a8090c4edd5eae"} Jan 26 15:58:35 crc kubenswrapper[4880]: I0126 15:58:35.687359 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:35 crc kubenswrapper[4880]: I0126 15:58:35.718604 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" podStartSLOduration=33.718575553 podStartE2EDuration="33.718575553s" podCreationTimestamp="2026-01-26 15:58:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:58:35.713166671 +0000 UTC m=+241.198895378" watchObservedRunningTime="2026-01-26 15:58:35.718575553 +0000 UTC m=+241.204304260" Jan 26 15:58:36 crc kubenswrapper[4880]: I0126 15:58:36.110914 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-754dc54bdd-qk4x6" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.517729 4880 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.518371 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e" gracePeriod=15 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.518546 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0" gracePeriod=15 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.518597 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a" gracePeriod=15 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.518600 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2" gracePeriod=15 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.518670 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49" gracePeriod=15 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519051 4880 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519306 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519323 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519334 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519340 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519347 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519353 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519361 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519367 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519375 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519382 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519390 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519398 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519413 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519420 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519539 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519549 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519557 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519565 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519573 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519585 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: E0126 15:58:38.519716 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519728 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.519811 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.521023 4880 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.521605 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.528960 4880 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661107 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661163 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661193 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661278 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661318 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661354 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661385 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.661430 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.707749 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709044 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709696 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0" exitCode=0 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709725 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49" exitCode=0 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709733 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2" exitCode=0 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709740 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a" exitCode=2 Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.709776 4880 scope.go:117] "RemoveContainer" containerID="fadc2d8a585bfeffb95ec98f3c79d7d404e3756b6ec1bd628dbbe88c300476f7" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.763174 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.763330 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.763945 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764042 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764070 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764100 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764142 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764159 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764186 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764221 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764274 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764253 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764002 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764330 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764344 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:38 crc kubenswrapper[4880]: I0126 15:58:38.764366 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.227350 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.228784 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.470134 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.470993 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.471562 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.720233 4880 generic.go:334] "Generic (PLEG): container finished" podID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" containerID="63d2b6b8ae3778f6df81ac62ba335b9fe541abff855ebf132a9ae7b652d9ea61" exitCode=0 Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.720336 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3f83e637-92b1-41a8-9df9-f1d5420ca1f8","Type":"ContainerDied","Data":"63d2b6b8ae3778f6df81ac62ba335b9fe541abff855ebf132a9ae7b652d9ea61"} Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.721145 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.721489 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.721998 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:39 crc kubenswrapper[4880]: I0126 15:58:39.725117 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.423856 4880 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.424623 4880 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.425080 4880 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.425363 4880 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.425631 4880 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.425668 4880 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.425840 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="200ms" Jan 26 15:58:40 crc kubenswrapper[4880]: E0126 15:58:40.641802 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="400ms" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.930349 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.931706 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.932761 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.933319 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.933865 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:40 crc kubenswrapper[4880]: I0126 15:58:40.934173 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.006417 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.006914 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.006970 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.006718 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.007278 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.007344 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.010744 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.011497 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.012021 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.012722 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.013039 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.044874 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="800ms" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.107836 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock\") pod \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108000 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir\") pod \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108109 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access\") pod \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\" (UID: \"3f83e637-92b1-41a8-9df9-f1d5420ca1f8\") " Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108179 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock" (OuterVolumeSpecName: "var-lock") pod "3f83e637-92b1-41a8-9df9-f1d5420ca1f8" (UID: "3f83e637-92b1-41a8-9df9-f1d5420ca1f8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108203 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3f83e637-92b1-41a8-9df9-f1d5420ca1f8" (UID: "3f83e637-92b1-41a8-9df9-f1d5420ca1f8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108567 4880 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108590 4880 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108599 4880 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108608 4880 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.108619 4880 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.115895 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3f83e637-92b1-41a8-9df9-f1d5420ca1f8" (UID: "3f83e637-92b1-41a8-9df9-f1d5420ca1f8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.210040 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f83e637-92b1-41a8-9df9-f1d5420ca1f8-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.741124 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.742102 4880 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e" exitCode=0 Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.742318 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.743274 4880 scope.go:117] "RemoveContainer" containerID="d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.745004 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3f83e637-92b1-41a8-9df9-f1d5420ca1f8","Type":"ContainerDied","Data":"3431f25d8b899d8c92b1c7350a56f3b25c0bac0d43a45e04e7325275296db547"} Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.745043 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3431f25d8b899d8c92b1c7350a56f3b25c0bac0d43a45e04e7325275296db547" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.745130 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.763674 4880 scope.go:117] "RemoveContainer" containerID="4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.765288 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.766020 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.766648 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.767184 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.767726 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.768095 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.769355 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.769706 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.784901 4880 scope.go:117] "RemoveContainer" containerID="33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.802804 4880 scope.go:117] "RemoveContainer" containerID="7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.820614 4880 scope.go:117] "RemoveContainer" containerID="bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.840086 4880 scope.go:117] "RemoveContainer" containerID="c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.845888 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="1.6s" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.870287 4880 scope.go:117] "RemoveContainer" containerID="d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.870996 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\": container with ID starting with d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0 not found: ID does not exist" containerID="d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.871117 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0"} err="failed to get container status \"d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\": rpc error: code = NotFound desc = could not find container \"d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0\": container with ID starting with d88d6e8f83a03d73a4fc907cf27737c404f25ed61f2d0455ea896b882e99b6d0 not found: ID does not exist" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.871242 4880 scope.go:117] "RemoveContainer" containerID="4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.871964 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\": container with ID starting with 4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49 not found: ID does not exist" containerID="4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.871992 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49"} err="failed to get container status \"4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\": rpc error: code = NotFound desc = could not find container \"4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49\": container with ID starting with 4137befee78ffc86d943b1db333bc71cc5dba5c7e1cd030969d9005f834cba49 not found: ID does not exist" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.872007 4880 scope.go:117] "RemoveContainer" containerID="33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.872332 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\": container with ID starting with 33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2 not found: ID does not exist" containerID="33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.872379 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2"} err="failed to get container status \"33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\": rpc error: code = NotFound desc = could not find container \"33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2\": container with ID starting with 33e28654cf6e3c8295e211015a2b6b57fb79576ee0c7ad932654a2022e77bad2 not found: ID does not exist" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.872413 4880 scope.go:117] "RemoveContainer" containerID="7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.872970 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\": container with ID starting with 7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a not found: ID does not exist" containerID="7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.873071 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a"} err="failed to get container status \"7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\": rpc error: code = NotFound desc = could not find container \"7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a\": container with ID starting with 7b086425eb75727f9c0756e6056e032b1babf60392c3ffb84167b77ccf10b60a not found: ID does not exist" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.873189 4880 scope.go:117] "RemoveContainer" containerID="bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.873703 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\": container with ID starting with bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e not found: ID does not exist" containerID="bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.873818 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e"} err="failed to get container status \"bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\": rpc error: code = NotFound desc = could not find container \"bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e\": container with ID starting with bc9b086de423cdba81f6a2078b70d450d1cfad9cebc8c16dd07f715ea503499e not found: ID does not exist" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.873928 4880 scope.go:117] "RemoveContainer" containerID="c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd" Jan 26 15:58:41 crc kubenswrapper[4880]: E0126 15:58:41.874592 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\": container with ID starting with c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd not found: ID does not exist" containerID="c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd" Jan 26 15:58:41 crc kubenswrapper[4880]: I0126 15:58:41.874683 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd"} err="failed to get container status \"c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\": rpc error: code = NotFound desc = could not find container \"c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd\": container with ID starting with c69256083e29ee0f74c30e9396eab789fb899203ef89fcf8eee8c2df6bf98fbd not found: ID does not exist" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.049735 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.050269 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.050647 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.051013 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.051413 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.051688 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.094256 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.094913 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.096175 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.097072 4880 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.097396 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.097763 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:42 crc kubenswrapper[4880]: I0126 15:58:42.220613 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 26 15:58:43 crc kubenswrapper[4880]: E0126 15:58:43.446961 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="3.2s" Jan 26 15:58:43 crc kubenswrapper[4880]: E0126 15:58:43.553275 4880 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.188:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:43 crc kubenswrapper[4880]: I0126 15:58:43.553995 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:43 crc kubenswrapper[4880]: W0126 15:58:43.581138 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-343e4734472ea1e55aa0aebaa9e099b008ee9f9bccb62c5e7028a619dcd6a37d WatchSource:0}: Error finding container 343e4734472ea1e55aa0aebaa9e099b008ee9f9bccb62c5e7028a619dcd6a37d: Status 404 returned error can't find the container with id 343e4734472ea1e55aa0aebaa9e099b008ee9f9bccb62c5e7028a619dcd6a37d Jan 26 15:58:43 crc kubenswrapper[4880]: E0126 15:58:43.584900 4880 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.188:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e53255f7d2436 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 15:58:43.584115766 +0000 UTC m=+249.069844473,LastTimestamp:2026-01-26 15:58:43.584115766 +0000 UTC m=+249.069844473,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 15:58:43 crc kubenswrapper[4880]: I0126 15:58:43.760855 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"343e4734472ea1e55aa0aebaa9e099b008ee9f9bccb62c5e7028a619dcd6a37d"} Jan 26 15:58:44 crc kubenswrapper[4880]: I0126 15:58:44.767091 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"218a85af1cab4e88214512dd244ae1f3b417ac3f48eb68044e844ba2088ae851"} Jan 26 15:58:44 crc kubenswrapper[4880]: I0126 15:58:44.767896 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:44 crc kubenswrapper[4880]: E0126 15:58:44.768029 4880 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.188:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:44 crc kubenswrapper[4880]: I0126 15:58:44.768078 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:44 crc kubenswrapper[4880]: I0126 15:58:44.768235 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:44 crc kubenswrapper[4880]: I0126 15:58:44.768400 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:45 crc kubenswrapper[4880]: E0126 15:58:45.775307 4880 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.188:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:58:46 crc kubenswrapper[4880]: I0126 15:58:46.217019 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:46 crc kubenswrapper[4880]: I0126 15:58:46.217641 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:46 crc kubenswrapper[4880]: I0126 15:58:46.217993 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:46 crc kubenswrapper[4880]: I0126 15:58:46.218284 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:46 crc kubenswrapper[4880]: E0126 15:58:46.648348 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="6.4s" Jan 26 15:58:51 crc kubenswrapper[4880]: E0126 15:58:51.397745 4880 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.188:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e53255f7d2436 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 15:58:43.584115766 +0000 UTC m=+249.069844473,LastTimestamp:2026-01-26 15:58:43.584115766 +0000 UTC m=+249.069844473,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 15:58:53 crc kubenswrapper[4880]: E0126 15:58:53.050330 4880 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.188:6443: connect: connection refused" interval="7s" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.213723 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.215021 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.216523 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.217027 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.217340 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.238874 4880 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.238978 4880 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:53 crc kubenswrapper[4880]: E0126 15:58:53.239593 4880 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.240378 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:53 crc kubenswrapper[4880]: W0126 15:58:53.350888 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-f96d39d9c75add532722bcad03ce9c0c14751a1f0fd31a6f667d8645a3a93fe6 WatchSource:0}: Error finding container f96d39d9c75add532722bcad03ce9c0c14751a1f0fd31a6f667d8645a3a93fe6: Status 404 returned error can't find the container with id f96d39d9c75add532722bcad03ce9c0c14751a1f0fd31a6f667d8645a3a93fe6 Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.832849 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.833177 4880 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19" exitCode=1 Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.833260 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19"} Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.834056 4880 scope.go:117] "RemoveContainer" containerID="250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.834524 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.835519 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.836227 4880 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.836676 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.836923 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.837156 4880 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="76432c11923588b5da8a35450de4b0045e895421a7d7abae1af6cf03ae57454f" exitCode=0 Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.837197 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"76432c11923588b5da8a35450de4b0045e895421a7d7abae1af6cf03ae57454f"} Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.837226 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f96d39d9c75add532722bcad03ce9c0c14751a1f0fd31a6f667d8645a3a93fe6"} Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.837580 4880 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.837607 4880 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:53 crc kubenswrapper[4880]: E0126 15:58:53.838223 4880 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.839180 4880 status_manager.go:851] "Failed to get status for pod" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" pod="openshift-marketplace/redhat-operators-qr9hz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-qr9hz\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.839677 4880 status_manager.go:851] "Failed to get status for pod" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.840253 4880 status_manager.go:851] "Failed to get status for pod" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" pod="openshift-marketplace/community-operators-t9qds" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-t9qds\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.840865 4880 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:53 crc kubenswrapper[4880]: I0126 15:58:53.841193 4880 status_manager.go:851] "Failed to get status for pod" podUID="7019dea6-6033-4097-9410-092fe22a8402" pod="openshift-marketplace/certified-operators-r6zxd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-r6zxd\": dial tcp 38.102.83.188:6443: connect: connection refused" Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.244789 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.385291 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.855108 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"45c62bcd22727847d7201238bd5f5a65b62d26553855b5322d9f38535e92034f"} Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.855456 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"208d7ca12fc60451609c795dce54d0c1e217e30a6cf16dcafd1128b376d7ddf5"} Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.855474 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"21c49459be381f5d8f8ce55bc5b5106469ec9dd74c4c2b311b742baaf3d8a5c6"} Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.971771 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 15:58:54 crc kubenswrapper[4880]: I0126 15:58:54.971841 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"474480a4f0bf2305153250a90e8afd09c5eda48b7f4a759f57f83ff5d6ee555c"} Jan 26 15:58:56 crc kubenswrapper[4880]: I0126 15:58:56.175289 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"815acc1fb83e3d281bea94f289c166f6f39309b7242d9520cdbd5535982dcb9a"} Jan 26 15:58:57 crc kubenswrapper[4880]: I0126 15:58:57.185611 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1f314febf9f7c1ffc414b586c36b8d52926677b2c7548e71b2391fecbaa2b016"} Jan 26 15:58:57 crc kubenswrapper[4880]: I0126 15:58:57.185947 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:57 crc kubenswrapper[4880]: I0126 15:58:57.185998 4880 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:57 crc kubenswrapper[4880]: I0126 15:58:57.186021 4880 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:58:58 crc kubenswrapper[4880]: I0126 15:58:58.247012 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:58 crc kubenswrapper[4880]: I0126 15:58:58.247292 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:58:58 crc kubenswrapper[4880]: I0126 15:58:58.315850 4880 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]log ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]etcd ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/priority-and-fairness-filter ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-apiextensions-informers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-apiextensions-controllers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/crd-informer-synced ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-system-namespaces-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/bootstrap-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/start-kube-aggregator-informers ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-registration-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-discovery-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]autoregister-completion ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-openapi-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 26 15:58:58 crc kubenswrapper[4880]: livez check failed Jan 26 15:58:58 crc kubenswrapper[4880]: I0126 15:58:58.316010 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 15:59:02 crc kubenswrapper[4880]: I0126 15:59:02.289199 4880 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:59:02 crc kubenswrapper[4880]: I0126 15:59:02.421805 4880 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3a1915fb-b20c-45eb-a4a6-8e833676b4d1" Jan 26 15:59:03 crc kubenswrapper[4880]: I0126 15:59:03.280484 4880 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:59:03 crc kubenswrapper[4880]: I0126 15:59:03.280530 4880 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0599f916-ef50-4fe0-ac05-8a41facb7a16" Jan 26 15:59:03 crc kubenswrapper[4880]: I0126 15:59:03.283788 4880 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3a1915fb-b20c-45eb-a4a6-8e833676b4d1" Jan 26 15:59:04 crc kubenswrapper[4880]: I0126 15:59:04.244982 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:59:04 crc kubenswrapper[4880]: I0126 15:59:04.245300 4880 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 15:59:04 crc kubenswrapper[4880]: I0126 15:59:04.246617 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 15:59:04 crc kubenswrapper[4880]: I0126 15:59:04.385860 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:59:11 crc kubenswrapper[4880]: I0126 15:59:11.660863 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.080322 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.145305 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.222290 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.442260 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.637778 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.846038 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 15:59:12 crc kubenswrapper[4880]: I0126 15:59:12.857148 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 15:59:13 crc kubenswrapper[4880]: I0126 15:59:13.237588 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 15:59:13 crc kubenswrapper[4880]: I0126 15:59:13.292815 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 15:59:13 crc kubenswrapper[4880]: I0126 15:59:13.389007 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 15:59:13 crc kubenswrapper[4880]: I0126 15:59:13.468860 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 15:59:13 crc kubenswrapper[4880]: I0126 15:59:13.726789 4880 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.160473 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.245227 4880 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.245305 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.263785 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.305052 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.334576 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.408193 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.484640 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.516358 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.590543 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.601227 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 15:59:14 crc kubenswrapper[4880]: I0126 15:59:14.738830 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.268475 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.287417 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.300612 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.311562 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.396221 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.578559 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.691069 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.810130 4880 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.815151 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.815243 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.822029 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.834292 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.834264659 podStartE2EDuration="13.834264659s" podCreationTimestamp="2026-01-26 15:59:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 15:59:15.834111275 +0000 UTC m=+281.319840002" watchObservedRunningTime="2026-01-26 15:59:15.834264659 +0000 UTC m=+281.319993366" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.841232 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.884286 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.937141 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.966820 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 15:59:15 crc kubenswrapper[4880]: I0126 15:59:15.992141 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.004489 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.021836 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.156363 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.175755 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.225535 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.245711 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.245950 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.271806 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.294857 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.349832 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.426993 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.471858 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.520829 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.577307 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.716406 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 15:59:16 crc kubenswrapper[4880]: I0126 15:59:16.939202 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.095361 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.099528 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.244569 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.338997 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.522154 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.554424 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.570282 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.615333 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.711707 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.713324 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.724505 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.742239 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.756502 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.806283 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.814485 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.873671 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.913859 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.943097 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.965582 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.972671 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 15:59:17 crc kubenswrapper[4880]: I0126 15:59:17.975745 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.023228 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.042942 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.060091 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.160180 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.181586 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.215258 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.248752 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.249227 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.255778 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.262118 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.404136 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.407222 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.418905 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.426109 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.533944 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.540050 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.551884 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.575724 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.659722 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.672023 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.752715 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.774083 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.796309 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.804998 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.825400 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.855256 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 15:59:18 crc kubenswrapper[4880]: I0126 15:59:18.858930 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:18.940926 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.040899 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.062037 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.477122 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.528901 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.634612 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.797617 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.797701 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.815212 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.817404 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.819625 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 15:59:19 crc kubenswrapper[4880]: I0126 15:59:19.925977 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.029733 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.063706 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.108587 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.234044 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.282059 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.331982 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.355098 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.384882 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.444311 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.444420 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.462675 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.471883 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.517958 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.519984 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.533401 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.552121 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.589506 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.669503 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.673653 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 15:59:20 crc kubenswrapper[4880]: I0126 15:59:20.951952 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.021488 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.028392 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.211526 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.262619 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.311407 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.343608 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.430546 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.586369 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.612846 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.625177 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.690783 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.708018 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.740563 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.790594 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.908467 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.908819 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.912143 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.912216 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.913568 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 15:59:21 crc kubenswrapper[4880]: I0126 15:59:21.966765 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.023287 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.059560 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.102905 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.161022 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.166106 4880 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.188399 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.204912 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.254031 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.272181 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.379293 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.428149 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.444259 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.551826 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.671905 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.672187 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.672319 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.705255 4880 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.747524 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.784512 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.855024 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 15:59:22 crc kubenswrapper[4880]: I0126 15:59:22.936348 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.114642 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.386539 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.386879 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.387613 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.387741 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.387898 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.484143 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.510532 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.608825 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.947314 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.947676 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.947903 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 15:59:23 crc kubenswrapper[4880]: I0126 15:59:23.948047 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.065795 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.067823 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.090351 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.106814 4880 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.107175 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://218a85af1cab4e88214512dd244ae1f3b417ac3f48eb68044e844ba2088ae851" gracePeriod=5 Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.213299 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.245255 4880 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.245390 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.245495 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.246166 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"474480a4f0bf2305153250a90e8afd09c5eda48b7f4a759f57f83ff5d6ee555c"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.246302 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://474480a4f0bf2305153250a90e8afd09c5eda48b7f4a759f57f83ff5d6ee555c" gracePeriod=30 Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.274730 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.308421 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.312468 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.331794 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.447115 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.509825 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.620152 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.632519 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.640578 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.672889 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.807710 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.871179 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.892919 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.961752 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.985261 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 15:59:24 crc kubenswrapper[4880]: I0126 15:59:24.987619 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.063922 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.139240 4880 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.176514 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.213987 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.324459 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.416822 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.452249 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.518418 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.634120 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.826124 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.855507 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.861721 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 15:59:25 crc kubenswrapper[4880]: I0126 15:59:25.912788 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.032894 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.111180 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.120809 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.256154 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.256356 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.292777 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.654488 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.669732 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.840927 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.876578 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.905172 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.910899 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.929075 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.930561 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 15:59:26 crc kubenswrapper[4880]: I0126 15:59:26.951042 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.081197 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.142033 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.146785 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.208548 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.236562 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.251466 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.251511 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.269468 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.417205 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.439002 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.447395 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.721659 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.780847 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 15:59:27 crc kubenswrapper[4880]: I0126 15:59:27.990499 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 15:59:28 crc kubenswrapper[4880]: I0126 15:59:28.071412 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 15:59:28 crc kubenswrapper[4880]: I0126 15:59:28.084134 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 15:59:28 crc kubenswrapper[4880]: I0126 15:59:28.364527 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 15:59:28 crc kubenswrapper[4880]: I0126 15:59:28.719161 4880 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.517480 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.517546 4880 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="218a85af1cab4e88214512dd244ae1f3b417ac3f48eb68044e844ba2088ae851" exitCode=137 Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.696451 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.696592 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.831979 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832538 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832703 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832828 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832168 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832642 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.832769 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.833328 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.833452 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.833736 4880 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.833829 4880 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.833968 4880 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.834060 4880 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.842672 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 15:59:29 crc kubenswrapper[4880]: I0126 15:59:29.935797 4880 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:30 crc kubenswrapper[4880]: I0126 15:59:30.223278 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 26 15:59:30 crc kubenswrapper[4880]: I0126 15:59:30.528088 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 15:59:30 crc kubenswrapper[4880]: I0126 15:59:30.528202 4880 scope.go:117] "RemoveContainer" containerID="218a85af1cab4e88214512dd244ae1f3b417ac3f48eb68044e844ba2088ae851" Jan 26 15:59:30 crc kubenswrapper[4880]: I0126 15:59:30.528268 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 15:59:35 crc kubenswrapper[4880]: I0126 15:59:35.840253 4880 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 26 15:59:43 crc kubenswrapper[4880]: I0126 15:59:43.218653 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.101402 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.104281 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t9qds" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="registry-server" containerID="cri-o://a02f9f7f40969bfc503b2362a64469cd69d300a0d9be53bbed04741e50a7a8d6" gracePeriod=2 Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.313616 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.314194 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r6zxd" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="registry-server" containerID="cri-o://8c21ea9d2eb01a3d364d798ebf077f5569b4e357b18e834c617f428cdc94185f" gracePeriod=2 Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.751589 4880 generic.go:334] "Generic (PLEG): container finished" podID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerID="a02f9f7f40969bfc503b2362a64469cd69d300a0d9be53bbed04741e50a7a8d6" exitCode=0 Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.751737 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerDied","Data":"a02f9f7f40969bfc503b2362a64469cd69d300a0d9be53bbed04741e50a7a8d6"} Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.755744 4880 generic.go:334] "Generic (PLEG): container finished" podID="7019dea6-6033-4097-9410-092fe22a8402" containerID="8c21ea9d2eb01a3d364d798ebf077f5569b4e357b18e834c617f428cdc94185f" exitCode=0 Jan 26 15:59:52 crc kubenswrapper[4880]: I0126 15:59:52.755829 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerDied","Data":"8c21ea9d2eb01a3d364d798ebf077f5569b4e357b18e834c617f428cdc94185f"} Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.158824 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.360614 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities\") pod \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.360746 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjdcs\" (UniqueName: \"kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs\") pod \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.360817 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content\") pod \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\" (UID: \"0f25591a-acee-4cc9-98b5-ca45d47ef1bf\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.362339 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities" (OuterVolumeSpecName: "utilities") pod "0f25591a-acee-4cc9-98b5-ca45d47ef1bf" (UID: "0f25591a-acee-4cc9-98b5-ca45d47ef1bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.371073 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs" (OuterVolumeSpecName: "kube-api-access-cjdcs") pod "0f25591a-acee-4cc9-98b5-ca45d47ef1bf" (UID: "0f25591a-acee-4cc9-98b5-ca45d47ef1bf"). InnerVolumeSpecName "kube-api-access-cjdcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.402296 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.437248 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f25591a-acee-4cc9-98b5-ca45d47ef1bf" (UID: "0f25591a-acee-4cc9-98b5-ca45d47ef1bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.462778 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjdcs\" (UniqueName: \"kubernetes.io/projected/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-kube-api-access-cjdcs\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.462840 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.462849 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f25591a-acee-4cc9-98b5-ca45d47ef1bf-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.563544 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvxx6\" (UniqueName: \"kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6\") pod \"7019dea6-6033-4097-9410-092fe22a8402\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.563654 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content\") pod \"7019dea6-6033-4097-9410-092fe22a8402\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.563701 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities\") pod \"7019dea6-6033-4097-9410-092fe22a8402\" (UID: \"7019dea6-6033-4097-9410-092fe22a8402\") " Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.564952 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities" (OuterVolumeSpecName: "utilities") pod "7019dea6-6033-4097-9410-092fe22a8402" (UID: "7019dea6-6033-4097-9410-092fe22a8402"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.566873 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6" (OuterVolumeSpecName: "kube-api-access-dvxx6") pod "7019dea6-6033-4097-9410-092fe22a8402" (UID: "7019dea6-6033-4097-9410-092fe22a8402"). InnerVolumeSpecName "kube-api-access-dvxx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.619189 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7019dea6-6033-4097-9410-092fe22a8402" (UID: "7019dea6-6033-4097-9410-092fe22a8402"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.665144 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvxx6\" (UniqueName: \"kubernetes.io/projected/7019dea6-6033-4097-9410-092fe22a8402-kube-api-access-dvxx6\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.665199 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.665213 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7019dea6-6033-4097-9410-092fe22a8402-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.765905 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r6zxd" event={"ID":"7019dea6-6033-4097-9410-092fe22a8402","Type":"ContainerDied","Data":"baf400fe29a08710b74f11311d99f507d2d82e59354895c536785430dd8cf41e"} Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.766289 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r6zxd" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.766408 4880 scope.go:117] "RemoveContainer" containerID="8c21ea9d2eb01a3d364d798ebf077f5569b4e357b18e834c617f428cdc94185f" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.769559 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9qds" event={"ID":"0f25591a-acee-4cc9-98b5-ca45d47ef1bf","Type":"ContainerDied","Data":"5c8f20307c46ae6432f5adebbfa31b9b80e07a56e43541b298ce066b4f93197a"} Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.769644 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9qds" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.796725 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.798677 4880 scope.go:117] "RemoveContainer" containerID="ad7cff0409979eda239a6ebfe4c93d18ec5967d61b2cb0b8146a9596b95cb2ab" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.802103 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r6zxd"] Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.811431 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.827862 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t9qds"] Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.830687 4880 scope.go:117] "RemoveContainer" containerID="250d2d8746f26b20afce5879f982c505109cc6c87c62b1703ca170bb8e5c7bcc" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.847786 4880 scope.go:117] "RemoveContainer" containerID="a02f9f7f40969bfc503b2362a64469cd69d300a0d9be53bbed04741e50a7a8d6" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.862565 4880 scope.go:117] "RemoveContainer" containerID="60a5d5e0206253a2fcec31676d6e1f69f841fb8a61805cf0d72bc0b0fabb667c" Jan 26 15:59:53 crc kubenswrapper[4880]: I0126 15:59:53.882317 4880 scope.go:117] "RemoveContainer" containerID="a2cfea7242b4033b4ee6d62088217819b86613d2c676ff386f0d1e23f0b37dec" Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.222750 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" path="/var/lib/kubelet/pods/0f25591a-acee-4cc9-98b5-ca45d47ef1bf/volumes" Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.223698 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7019dea6-6033-4097-9410-092fe22a8402" path="/var/lib/kubelet/pods/7019dea6-6033-4097-9410-092fe22a8402/volumes" Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.894806 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.896688 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.896745 4880 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="474480a4f0bf2305153250a90e8afd09c5eda48b7f4a759f57f83ff5d6ee555c" exitCode=137 Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.896907 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"474480a4f0bf2305153250a90e8afd09c5eda48b7f4a759f57f83ff5d6ee555c"} Jan 26 15:59:54 crc kubenswrapper[4880]: I0126 15:59:54.897093 4880 scope.go:117] "RemoveContainer" containerID="250e7ab3de4e83d888b512e14910a7e816082af03ada263d1d64538aab6faa19" Jan 26 15:59:55 crc kubenswrapper[4880]: I0126 15:59:55.912234 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 15:59:55 crc kubenswrapper[4880]: I0126 15:59:55.914081 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"42aafc445d4ac8bcea429fc76d4ef1c4b5ad81b0a8b0d46f89a332a79249d115"} Jan 26 15:59:58 crc kubenswrapper[4880]: I0126 15:59:58.382768 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 16:00:04 crc kubenswrapper[4880]: I0126 16:00:04.245670 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:00:04 crc kubenswrapper[4880]: I0126 16:00:04.251288 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:00:04 crc kubenswrapper[4880]: I0126 16:00:04.385843 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:00:05 crc kubenswrapper[4880]: I0126 16:00:05.975529 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.972962 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms"] Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973622 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="extract-content" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973643 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="extract-content" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973665 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973674 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973689 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="extract-content" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973698 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="extract-content" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973710 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="extract-utilities" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973717 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="extract-utilities" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973734 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="extract-utilities" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973741 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="extract-utilities" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973755 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" containerName="installer" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973762 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" containerName="installer" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973777 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973784 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:00:11 crc kubenswrapper[4880]: E0126 16:00:11.973797 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973804 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973958 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f83e637-92b1-41a8-9df9-f1d5420ca1f8" containerName="installer" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973978 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="7019dea6-6033-4097-9410-092fe22a8402" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973987 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.973999 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f25591a-acee-4cc9-98b5-ca45d47ef1bf" containerName="registry-server" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.974620 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.977652 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.977698 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 16:00:11 crc kubenswrapper[4880]: I0126 16:00:11.990460 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms"] Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.040398 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.040702 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" containerID="cri-o://30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719" gracePeriod=30 Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.044849 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.045155 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" podUID="125185c6-2155-4eef-ba91-e52ac6347421" containerName="route-controller-manager" containerID="cri-o://6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50" gracePeriod=30 Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.092296 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.092419 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtr9b\" (UniqueName: \"kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.092533 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.193254 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.193338 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.193385 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtr9b\" (UniqueName: \"kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.195410 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.209525 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.226996 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtr9b\" (UniqueName: \"kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b\") pod \"collect-profiles-29490720-cf7ms\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.301010 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.466691 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.599660 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert\") pod \"125185c6-2155-4eef-ba91-e52ac6347421\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.599739 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca\") pod \"125185c6-2155-4eef-ba91-e52ac6347421\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.599769 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84cd2\" (UniqueName: \"kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2\") pod \"125185c6-2155-4eef-ba91-e52ac6347421\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.599874 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config\") pod \"125185c6-2155-4eef-ba91-e52ac6347421\" (UID: \"125185c6-2155-4eef-ba91-e52ac6347421\") " Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.601141 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config" (OuterVolumeSpecName: "config") pod "125185c6-2155-4eef-ba91-e52ac6347421" (UID: "125185c6-2155-4eef-ba91-e52ac6347421"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.601284 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca" (OuterVolumeSpecName: "client-ca") pod "125185c6-2155-4eef-ba91-e52ac6347421" (UID: "125185c6-2155-4eef-ba91-e52ac6347421"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.605425 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "125185c6-2155-4eef-ba91-e52ac6347421" (UID: "125185c6-2155-4eef-ba91-e52ac6347421"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.606112 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2" (OuterVolumeSpecName: "kube-api-access-84cd2") pod "125185c6-2155-4eef-ba91-e52ac6347421" (UID: "125185c6-2155-4eef-ba91-e52ac6347421"). InnerVolumeSpecName "kube-api-access-84cd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.701114 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.701168 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/125185c6-2155-4eef-ba91-e52ac6347421-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.701190 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84cd2\" (UniqueName: \"kubernetes.io/projected/125185c6-2155-4eef-ba91-e52ac6347421-kube-api-access-84cd2\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.701218 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/125185c6-2155-4eef-ba91-e52ac6347421-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.851350 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms"] Jan 26 16:00:12 crc kubenswrapper[4880]: I0126 16:00:12.932917 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.101793 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" event={"ID":"c64c56d4-e01c-4b8d-8270-d0c6445d90e4","Type":"ContainerStarted","Data":"8b41d398fee6b64437b0b56c4673f83f7b989f0251ff85627774bbab0e249de6"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.101852 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" event={"ID":"c64c56d4-e01c-4b8d-8270-d0c6445d90e4","Type":"ContainerStarted","Data":"da4acf74749e873440cfd32b61638e7eaf0c940dddb004421981885cd6ca85c4"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104608 4880 generic.go:334] "Generic (PLEG): container finished" podID="125185c6-2155-4eef-ba91-e52ac6347421" containerID="6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50" exitCode=0 Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104702 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104692 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" event={"ID":"125185c6-2155-4eef-ba91-e52ac6347421","Type":"ContainerDied","Data":"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104631 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config\") pod \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104784 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b" event={"ID":"125185c6-2155-4eef-ba91-e52ac6347421","Type":"ContainerDied","Data":"0b09db30290395a7868f77de9356ac8bf83154a1b8331cd00f1c3844bb6025c3"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104812 4880 scope.go:117] "RemoveContainer" containerID="6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104893 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca\") pod \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.104941 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4l5w\" (UniqueName: \"kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w\") pod \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.105044 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert\") pod \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.105121 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles\") pod \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\" (UID: \"875a4230-6bbb-4b42-b959-33ceeef0e3c4\") " Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.105733 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca" (OuterVolumeSpecName: "client-ca") pod "875a4230-6bbb-4b42-b959-33ceeef0e3c4" (UID: "875a4230-6bbb-4b42-b959-33ceeef0e3c4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.105815 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config" (OuterVolumeSpecName: "config") pod "875a4230-6bbb-4b42-b959-33ceeef0e3c4" (UID: "875a4230-6bbb-4b42-b959-33ceeef0e3c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.106410 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "875a4230-6bbb-4b42-b959-33ceeef0e3c4" (UID: "875a4230-6bbb-4b42-b959-33ceeef0e3c4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.111470 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w" (OuterVolumeSpecName: "kube-api-access-s4l5w") pod "875a4230-6bbb-4b42-b959-33ceeef0e3c4" (UID: "875a4230-6bbb-4b42-b959-33ceeef0e3c4"). InnerVolumeSpecName "kube-api-access-s4l5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.112127 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "875a4230-6bbb-4b42-b959-33ceeef0e3c4" (UID: "875a4230-6bbb-4b42-b959-33ceeef0e3c4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.113425 4880 generic.go:334] "Generic (PLEG): container finished" podID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerID="30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719" exitCode=0 Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.113494 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" event={"ID":"875a4230-6bbb-4b42-b959-33ceeef0e3c4","Type":"ContainerDied","Data":"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.113555 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" event={"ID":"875a4230-6bbb-4b42-b959-33ceeef0e3c4","Type":"ContainerDied","Data":"63a469ae102c82a2b87493b69b853c5821973a52de2fe56c84bccb545a2e1f3e"} Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.113572 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4xvtl" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.145885 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" podStartSLOduration=2.145842213 podStartE2EDuration="2.145842213s" podCreationTimestamp="2026-01-26 16:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:13.121668696 +0000 UTC m=+338.607397393" watchObservedRunningTime="2026-01-26 16:00:13.145842213 +0000 UTC m=+338.631570920" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.147374 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.148783 4880 scope.go:117] "RemoveContainer" containerID="6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50" Jan 26 16:00:13 crc kubenswrapper[4880]: E0126 16:00:13.149769 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50\": container with ID starting with 6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50 not found: ID does not exist" containerID="6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.149813 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50"} err="failed to get container status \"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50\": rpc error: code = NotFound desc = could not find container \"6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50\": container with ID starting with 6b59b2654c8ae29d4d19facaba66109d8221fdcbe555d263356dc46280d82e50 not found: ID does not exist" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.149837 4880 scope.go:117] "RemoveContainer" containerID="30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.152068 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lmp9b"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.162484 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.166696 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4xvtl"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.198120 4880 scope.go:117] "RemoveContainer" containerID="30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719" Jan 26 16:00:13 crc kubenswrapper[4880]: E0126 16:00:13.198764 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719\": container with ID starting with 30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719 not found: ID does not exist" containerID="30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.199216 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719"} err="failed to get container status \"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719\": rpc error: code = NotFound desc = could not find container \"30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719\": container with ID starting with 30afe923b0a1b26aa810aef7ae109f63cfe3fe480bb49298deda2d66a50d3719 not found: ID does not exist" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.207212 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.207566 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.207692 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/875a4230-6bbb-4b42-b959-33ceeef0e3c4-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.207800 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4l5w\" (UniqueName: \"kubernetes.io/projected/875a4230-6bbb-4b42-b959-33ceeef0e3c4-kube-api-access-s4l5w\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.207895 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/875a4230-6bbb-4b42-b959-33ceeef0e3c4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.669106 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:13 crc kubenswrapper[4880]: E0126 16:00:13.669463 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125185c6-2155-4eef-ba91-e52ac6347421" containerName="route-controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.669477 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="125185c6-2155-4eef-ba91-e52ac6347421" containerName="route-controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: E0126 16:00:13.669490 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.669496 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.669591 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" containerName="controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.669606 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="125185c6-2155-4eef-ba91-e52ac6347421" containerName="route-controller-manager" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.670137 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.673278 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.674396 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.674400 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.674400 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.675476 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.676901 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.675805 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.675856 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.676420 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.677579 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.677744 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.679516 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.684253 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.686681 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.687833 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.692054 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.698644 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.814278 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.814808 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tgdw\" (UniqueName: \"kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.815009 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.815236 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.815499 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.815724 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.815900 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.816067 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swxph\" (UniqueName: \"kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.816222 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918031 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918105 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918134 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918177 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swxph\" (UniqueName: \"kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918226 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918269 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tgdw\" (UniqueName: \"kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918315 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918341 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.918414 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.920098 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.921214 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.921508 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.921512 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.923017 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.925212 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:13 crc kubenswrapper[4880]: I0126 16:00:13.931967 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.219066 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tgdw\" (UniqueName: \"kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw\") pod \"route-controller-manager-78c9545596-cvkzg\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.236803 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swxph\" (UniqueName: \"kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph\") pod \"controller-manager-789d9f7b94-zk4ts\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.237326 4880 generic.go:334] "Generic (PLEG): container finished" podID="c64c56d4-e01c-4b8d-8270-d0c6445d90e4" containerID="8b41d398fee6b64437b0b56c4673f83f7b989f0251ff85627774bbab0e249de6" exitCode=0 Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.237592 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="125185c6-2155-4eef-ba91-e52ac6347421" path="/var/lib/kubelet/pods/125185c6-2155-4eef-ba91-e52ac6347421/volumes" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.238574 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875a4230-6bbb-4b42-b959-33ceeef0e3c4" path="/var/lib/kubelet/pods/875a4230-6bbb-4b42-b959-33ceeef0e3c4/volumes" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.239036 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" event={"ID":"c64c56d4-e01c-4b8d-8270-d0c6445d90e4","Type":"ContainerDied","Data":"8b41d398fee6b64437b0b56c4673f83f7b989f0251ff85627774bbab0e249de6"} Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.294560 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.306486 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.706302 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:00:14 crc kubenswrapper[4880]: I0126 16:00:14.966337 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:14 crc kubenswrapper[4880]: W0126 16:00:14.971993 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75bff2be_61c4_4f41_a422_d19d73159763.slice/crio-4770f444ffcb68d5ef919a58784e2a1820cec18bbe2f760a19fceee71531d5c4 WatchSource:0}: Error finding container 4770f444ffcb68d5ef919a58784e2a1820cec18bbe2f760a19fceee71531d5c4: Status 404 returned error can't find the container with id 4770f444ffcb68d5ef919a58784e2a1820cec18bbe2f760a19fceee71531d5c4 Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.267468 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" event={"ID":"75bff2be-61c4-4f41-a422-d19d73159763","Type":"ContainerStarted","Data":"bd0b0d3f179d0cb4bf981c2a17e795d981cb6e5944c202026dab4cb8814acd96"} Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.267553 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" event={"ID":"75bff2be-61c4-4f41-a422-d19d73159763","Type":"ContainerStarted","Data":"4770f444ffcb68d5ef919a58784e2a1820cec18bbe2f760a19fceee71531d5c4"} Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.268275 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.271817 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" event={"ID":"a70cd4a9-e2ee-456d-a032-9fb0607dc91a","Type":"ContainerStarted","Data":"e4e636a9ad1440b1fd928900036986c6d7ce6d37863a7708776d980ac9874f56"} Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.271937 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.271966 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" event={"ID":"a70cd4a9-e2ee-456d-a032-9fb0607dc91a","Type":"ContainerStarted","Data":"e50286eead330331a71ef1742053652f57f48204757d3dfb7201528d0e82ff95"} Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.282837 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.302084 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" podStartSLOduration=3.3020397089999998 podStartE2EDuration="3.302039709s" podCreationTimestamp="2026-01-26 16:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:15.29541278 +0000 UTC m=+340.781141487" watchObservedRunningTime="2026-01-26 16:00:15.302039709 +0000 UTC m=+340.787768416" Jan 26 16:00:15 crc kubenswrapper[4880]: I0126 16:00:15.439033 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" podStartSLOduration=3.43894821 podStartE2EDuration="3.43894821s" podCreationTimestamp="2026-01-26 16:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:15.412357148 +0000 UTC m=+340.898085865" watchObservedRunningTime="2026-01-26 16:00:15.43894821 +0000 UTC m=+340.924676927" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.398587 4880 patch_prober.go:28] interesting pod/route-controller-manager-78c9545596-cvkzg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.482410 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" event={"ID":"c64c56d4-e01c-4b8d-8270-d0c6445d90e4","Type":"ContainerDied","Data":"da4acf74749e873440cfd32b61638e7eaf0c940dddb004421981885cd6ca85c4"} Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.482527 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da4acf74749e873440cfd32b61638e7eaf0c940dddb004421981885cd6ca85c4" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.483576 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.525297 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume\") pod \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.526048 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtr9b\" (UniqueName: \"kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b\") pod \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.526343 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume\") pod \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\" (UID: \"c64c56d4-e01c-4b8d-8270-d0c6445d90e4\") " Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.656791 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume" (OuterVolumeSpecName: "config-volume") pod "c64c56d4-e01c-4b8d-8270-d0c6445d90e4" (UID: "c64c56d4-e01c-4b8d-8270-d0c6445d90e4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.457572 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" podUID="75bff2be-61c4-4f41-a422-d19d73159763" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.668603 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c64c56d4-e01c-4b8d-8270-d0c6445d90e4" (UID: "c64c56d4-e01c-4b8d-8270-d0c6445d90e4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.678664 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b" (OuterVolumeSpecName: "kube-api-access-jtr9b") pod "c64c56d4-e01c-4b8d-8270-d0c6445d90e4" (UID: "c64c56d4-e01c-4b8d-8270-d0c6445d90e4"). InnerVolumeSpecName "kube-api-access-jtr9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.723621 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.768388 4880 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.768468 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtr9b\" (UniqueName: \"kubernetes.io/projected/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-kube-api-access-jtr9b\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:16 crc kubenswrapper[4880]: I0126 16:00:16.768485 4880 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c64c56d4-e01c-4b8d-8270-d0c6445d90e4-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:17 crc kubenswrapper[4880]: I0126 16:00:17.485628 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490720-cf7ms" Jan 26 16:00:32 crc kubenswrapper[4880]: I0126 16:00:32.395261 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:32 crc kubenswrapper[4880]: I0126 16:00:32.396163 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" podUID="75bff2be-61c4-4f41-a422-d19d73159763" containerName="route-controller-manager" containerID="cri-o://bd0b0d3f179d0cb4bf981c2a17e795d981cb6e5944c202026dab4cb8814acd96" gracePeriod=30 Jan 26 16:00:32 crc kubenswrapper[4880]: I0126 16:00:32.821478 4880 generic.go:334] "Generic (PLEG): container finished" podID="75bff2be-61c4-4f41-a422-d19d73159763" containerID="bd0b0d3f179d0cb4bf981c2a17e795d981cb6e5944c202026dab4cb8814acd96" exitCode=0 Jan 26 16:00:32 crc kubenswrapper[4880]: I0126 16:00:32.821542 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" event={"ID":"75bff2be-61c4-4f41-a422-d19d73159763","Type":"ContainerDied","Data":"bd0b0d3f179d0cb4bf981c2a17e795d981cb6e5944c202026dab4cb8814acd96"} Jan 26 16:00:32 crc kubenswrapper[4880]: I0126 16:00:32.880762 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.060427 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca\") pod \"75bff2be-61c4-4f41-a422-d19d73159763\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.060554 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert\") pod \"75bff2be-61c4-4f41-a422-d19d73159763\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.060593 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tgdw\" (UniqueName: \"kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw\") pod \"75bff2be-61c4-4f41-a422-d19d73159763\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.060652 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config\") pod \"75bff2be-61c4-4f41-a422-d19d73159763\" (UID: \"75bff2be-61c4-4f41-a422-d19d73159763\") " Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.061557 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca" (OuterVolumeSpecName: "client-ca") pod "75bff2be-61c4-4f41-a422-d19d73159763" (UID: "75bff2be-61c4-4f41-a422-d19d73159763"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.061569 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config" (OuterVolumeSpecName: "config") pod "75bff2be-61c4-4f41-a422-d19d73159763" (UID: "75bff2be-61c4-4f41-a422-d19d73159763"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.066521 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw" (OuterVolumeSpecName: "kube-api-access-5tgdw") pod "75bff2be-61c4-4f41-a422-d19d73159763" (UID: "75bff2be-61c4-4f41-a422-d19d73159763"). InnerVolumeSpecName "kube-api-access-5tgdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.066852 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "75bff2be-61c4-4f41-a422-d19d73159763" (UID: "75bff2be-61c4-4f41-a422-d19d73159763"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.162835 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.162920 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75bff2be-61c4-4f41-a422-d19d73159763-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.162945 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tgdw\" (UniqueName: \"kubernetes.io/projected/75bff2be-61c4-4f41-a422-d19d73159763-kube-api-access-5tgdw\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.162966 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75bff2be-61c4-4f41-a422-d19d73159763-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.760230 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr"] Jan 26 16:00:33 crc kubenswrapper[4880]: E0126 16:00:33.760525 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75bff2be-61c4-4f41-a422-d19d73159763" containerName="route-controller-manager" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.760563 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="75bff2be-61c4-4f41-a422-d19d73159763" containerName="route-controller-manager" Jan 26 16:00:33 crc kubenswrapper[4880]: E0126 16:00:33.760600 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64c56d4-e01c-4b8d-8270-d0c6445d90e4" containerName="collect-profiles" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.760607 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64c56d4-e01c-4b8d-8270-d0c6445d90e4" containerName="collect-profiles" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.760750 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="75bff2be-61c4-4f41-a422-d19d73159763" containerName="route-controller-manager" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.760771 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64c56d4-e01c-4b8d-8270-d0c6445d90e4" containerName="collect-profiles" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.761270 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.778742 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr"] Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.828446 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" event={"ID":"75bff2be-61c4-4f41-a422-d19d73159763","Type":"ContainerDied","Data":"4770f444ffcb68d5ef919a58784e2a1820cec18bbe2f760a19fceee71531d5c4"} Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.828520 4880 scope.go:117] "RemoveContainer" containerID="bd0b0d3f179d0cb4bf981c2a17e795d981cb6e5944c202026dab4cb8814acd96" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.828528 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.866856 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.870840 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78c9545596-cvkzg"] Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.871243 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xf4l\" (UniqueName: \"kubernetes.io/projected/d435ee80-d9f3-4c21-b175-c678b308f4e9-kube-api-access-8xf4l\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.871321 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-config\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.871541 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-client-ca\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.871639 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d435ee80-d9f3-4c21-b175-c678b308f4e9-serving-cert\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.972682 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-client-ca\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.972795 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d435ee80-d9f3-4c21-b175-c678b308f4e9-serving-cert\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.972866 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xf4l\" (UniqueName: \"kubernetes.io/projected/d435ee80-d9f3-4c21-b175-c678b308f4e9-kube-api-access-8xf4l\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.972955 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-config\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.974138 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-client-ca\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.974583 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d435ee80-d9f3-4c21-b175-c678b308f4e9-config\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.983987 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d435ee80-d9f3-4c21-b175-c678b308f4e9-serving-cert\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:33 crc kubenswrapper[4880]: I0126 16:00:33.995379 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xf4l\" (UniqueName: \"kubernetes.io/projected/d435ee80-d9f3-4c21-b175-c678b308f4e9-kube-api-access-8xf4l\") pod \"route-controller-manager-594c99b869-5z7dr\" (UID: \"d435ee80-d9f3-4c21-b175-c678b308f4e9\") " pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.102678 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.224035 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75bff2be-61c4-4f41-a422-d19d73159763" path="/var/lib/kubelet/pods/75bff2be-61c4-4f41-a422-d19d73159763/volumes" Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.541534 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr"] Jan 26 16:00:34 crc kubenswrapper[4880]: W0126 16:00:34.548232 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd435ee80_d9f3_4c21_b175_c678b308f4e9.slice/crio-9642f629fe727e21a28efe65ed58a5da1144a4db566c4fb4a439d79b630f2298 WatchSource:0}: Error finding container 9642f629fe727e21a28efe65ed58a5da1144a4db566c4fb4a439d79b630f2298: Status 404 returned error can't find the container with id 9642f629fe727e21a28efe65ed58a5da1144a4db566c4fb4a439d79b630f2298 Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.834905 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" event={"ID":"d435ee80-d9f3-4c21-b175-c678b308f4e9","Type":"ContainerStarted","Data":"ea7b10b0bd48ed8df25182dafce540b6ba49776845f0eb71632cd478af8362cc"} Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.834954 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" event={"ID":"d435ee80-d9f3-4c21-b175-c678b308f4e9","Type":"ContainerStarted","Data":"9642f629fe727e21a28efe65ed58a5da1144a4db566c4fb4a439d79b630f2298"} Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.835385 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:34 crc kubenswrapper[4880]: I0126 16:00:34.852754 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" podStartSLOduration=2.852722509 podStartE2EDuration="2.852722509s" podCreationTimestamp="2026-01-26 16:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:34.849391098 +0000 UTC m=+360.335119805" watchObservedRunningTime="2026-01-26 16:00:34.852722509 +0000 UTC m=+360.338451216" Jan 26 16:00:35 crc kubenswrapper[4880]: I0126 16:00:35.457131 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-594c99b869-5z7dr" Jan 26 16:00:37 crc kubenswrapper[4880]: I0126 16:00:37.982340 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9gv9"] Jan 26 16:00:37 crc kubenswrapper[4880]: I0126 16:00:37.984495 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.020336 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9gv9"] Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.192315 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-bound-sa-token\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.192724 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.192884 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-tls\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.193144 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-trusted-ca\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.193282 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.193417 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-certificates\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.193590 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.193755 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pbl4\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-kube-api-access-8pbl4\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.224038 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295070 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-trusted-ca\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295489 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295531 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295558 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-certificates\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295601 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pbl4\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-kube-api-access-8pbl4\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295654 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-bound-sa-token\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.295708 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-tls\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.297200 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-trusted-ca\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.298714 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.303585 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-tls\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.304603 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.305806 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-registry-certificates\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.318343 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pbl4\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-kube-api-access-8pbl4\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.333059 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/985c7455-0f58-4ed7-a437-8e2c5f61e3d7-bound-sa-token\") pod \"image-registry-66df7c8f76-p9gv9\" (UID: \"985c7455-0f58-4ed7-a437-8e2c5f61e3d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:38 crc kubenswrapper[4880]: I0126 16:00:38.612190 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:39 crc kubenswrapper[4880]: I0126 16:00:39.038236 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-p9gv9"] Jan 26 16:00:39 crc kubenswrapper[4880]: W0126 16:00:39.048935 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985c7455_0f58_4ed7_a437_8e2c5f61e3d7.slice/crio-d68c96853d46210c227ce424d27c32b7b875a51f8cdaeb358538f04c21fc1e92 WatchSource:0}: Error finding container d68c96853d46210c227ce424d27c32b7b875a51f8cdaeb358538f04c21fc1e92: Status 404 returned error can't find the container with id d68c96853d46210c227ce424d27c32b7b875a51f8cdaeb358538f04c21fc1e92 Jan 26 16:00:39 crc kubenswrapper[4880]: I0126 16:00:39.940499 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" event={"ID":"985c7455-0f58-4ed7-a437-8e2c5f61e3d7","Type":"ContainerStarted","Data":"79f0ed07ef4cb3d35e70c9da268a55aaa3f61c417e3de624cd44eef1b7e27d91"} Jan 26 16:00:39 crc kubenswrapper[4880]: I0126 16:00:39.940848 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:39 crc kubenswrapper[4880]: I0126 16:00:39.940867 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" event={"ID":"985c7455-0f58-4ed7-a437-8e2c5f61e3d7","Type":"ContainerStarted","Data":"d68c96853d46210c227ce424d27c32b7b875a51f8cdaeb358538f04c21fc1e92"} Jan 26 16:00:39 crc kubenswrapper[4880]: I0126 16:00:39.962919 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" podStartSLOduration=2.962884311 podStartE2EDuration="2.962884311s" podCreationTimestamp="2026-01-26 16:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:39.957493464 +0000 UTC m=+365.443222181" watchObservedRunningTime="2026-01-26 16:00:39.962884311 +0000 UTC m=+365.448613018" Jan 26 16:00:40 crc kubenswrapper[4880]: I0126 16:00:40.699876 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:00:40 crc kubenswrapper[4880]: I0126 16:00:40.699955 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.567907 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.570712 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5d9pw" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="registry-server" containerID="cri-o://894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" gracePeriod=30 Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.590314 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.590735 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4xxrk" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="registry-server" containerID="cri-o://e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" gracePeriod=30 Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.610770 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.611077 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" containerID="cri-o://bb2d56aac038c8b94316087d7e36c62df80a79abf07acb027f55b010328daf6e" gracePeriod=30 Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.613203 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.613590 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r4dp4" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="registry-server" containerID="cri-o://2147c5413f25743ec2170f5006969eb3d9047ad21d57a3b4d7f87ee419bbf6a8" gracePeriod=30 Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.621693 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zlnn7"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.622395 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.625073 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.625305 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qr9hz" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="registry-server" containerID="cri-o://30c7c6d17d43b2796d289522c26623a2bf6197f2d1b56093a53c0a73db9c52af" gracePeriod=30 Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.641809 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zlnn7"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.773352 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0 is running failed: container process not found" containerID="894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.773714 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0 is running failed: container process not found" containerID="894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.774075 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0 is running failed: container process not found" containerID="894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.774121 4880 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-5d9pw" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="registry-server" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.797678 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.797751 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lr9\" (UniqueName: \"kubernetes.io/projected/c376a5ed-f9eb-4592-b6b5-2e021c59d978-kube-api-access-t8lr9\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.797843 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.898965 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.899067 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.899102 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lr9\" (UniqueName: \"kubernetes.io/projected/c376a5ed-f9eb-4592-b6b5-2e021c59d978-kube-api-access-t8lr9\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.906905 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.908705 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c376a5ed-f9eb-4592-b6b5-2e021c59d978-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.920980 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lr9\" (UniqueName: \"kubernetes.io/projected/c376a5ed-f9eb-4592-b6b5-2e021c59d978-kube-api-access-t8lr9\") pod \"marketplace-operator-79b997595-zlnn7\" (UID: \"c376a5ed-f9eb-4592-b6b5-2e021c59d978\") " pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: I0126 16:00:48.952641 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.988241 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149 is running failed: container process not found" containerID="e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.988667 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149 is running failed: container process not found" containerID="e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.988939 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149 is running failed: container process not found" containerID="e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:00:48 crc kubenswrapper[4880]: E0126 16:00:48.988989 4880 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-4xxrk" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="registry-server" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.011231 4880 generic.go:334] "Generic (PLEG): container finished" podID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerID="30c7c6d17d43b2796d289522c26623a2bf6197f2d1b56093a53c0a73db9c52af" exitCode=0 Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.011291 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerDied","Data":"30c7c6d17d43b2796d289522c26623a2bf6197f2d1b56093a53c0a73db9c52af"} Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.012518 4880 generic.go:334] "Generic (PLEG): container finished" podID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerID="bb2d56aac038c8b94316087d7e36c62df80a79abf07acb027f55b010328daf6e" exitCode=0 Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.012557 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" event={"ID":"c8af8728-cae3-44d4-9f71-79c69e8c5646","Type":"ContainerDied","Data":"bb2d56aac038c8b94316087d7e36c62df80a79abf07acb027f55b010328daf6e"} Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.014108 4880 generic.go:334] "Generic (PLEG): container finished" podID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerID="e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" exitCode=0 Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.014143 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerDied","Data":"e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149"} Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.022832 4880 generic.go:334] "Generic (PLEG): container finished" podID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerID="894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" exitCode=0 Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.022921 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerDied","Data":"894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0"} Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.029242 4880 generic.go:334] "Generic (PLEG): container finished" podID="60d83505-44fa-47d9-b05c-def5eda227a9" containerID="2147c5413f25743ec2170f5006969eb3d9047ad21d57a3b4d7f87ee419bbf6a8" exitCode=0 Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.029288 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerDied","Data":"2147c5413f25743ec2170f5006969eb3d9047ad21d57a3b4d7f87ee419bbf6a8"} Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.154352 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.306115 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities\") pod \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.306241 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmdfr\" (UniqueName: \"kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr\") pod \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.306278 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content\") pod \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\" (UID: \"ebb3fdcf-80cb-4293-9e7d-17a44072c91e\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.307741 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities" (OuterVolumeSpecName: "utilities") pod "ebb3fdcf-80cb-4293-9e7d-17a44072c91e" (UID: "ebb3fdcf-80cb-4293-9e7d-17a44072c91e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.313758 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr" (OuterVolumeSpecName: "kube-api-access-bmdfr") pod "ebb3fdcf-80cb-4293-9e7d-17a44072c91e" (UID: "ebb3fdcf-80cb-4293-9e7d-17a44072c91e"). InnerVolumeSpecName "kube-api-access-bmdfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.318137 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.331723 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.366331 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.397232 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebb3fdcf-80cb-4293-9e7d-17a44072c91e" (UID: "ebb3fdcf-80cb-4293-9e7d-17a44072c91e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.408411 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.408602 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmdfr\" (UniqueName: \"kubernetes.io/projected/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-kube-api-access-bmdfr\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.408624 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebb3fdcf-80cb-4293-9e7d-17a44072c91e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509103 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqd95\" (UniqueName: \"kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95\") pod \"4b678a2b-2601-4267-8f78-9b6262b8261a\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509707 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59ckh\" (UniqueName: \"kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh\") pod \"c8af8728-cae3-44d4-9f71-79c69e8c5646\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509785 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content\") pod \"60d83505-44fa-47d9-b05c-def5eda227a9\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509826 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics\") pod \"c8af8728-cae3-44d4-9f71-79c69e8c5646\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509849 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content\") pod \"4b678a2b-2601-4267-8f78-9b6262b8261a\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509954 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca\") pod \"c8af8728-cae3-44d4-9f71-79c69e8c5646\" (UID: \"c8af8728-cae3-44d4-9f71-79c69e8c5646\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.509977 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities\") pod \"4b678a2b-2601-4267-8f78-9b6262b8261a\" (UID: \"4b678a2b-2601-4267-8f78-9b6262b8261a\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.510001 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhhm7\" (UniqueName: \"kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7\") pod \"60d83505-44fa-47d9-b05c-def5eda227a9\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.510042 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities\") pod \"60d83505-44fa-47d9-b05c-def5eda227a9\" (UID: \"60d83505-44fa-47d9-b05c-def5eda227a9\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.510834 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c8af8728-cae3-44d4-9f71-79c69e8c5646" (UID: "c8af8728-cae3-44d4-9f71-79c69e8c5646"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.511030 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities" (OuterVolumeSpecName: "utilities") pod "4b678a2b-2601-4267-8f78-9b6262b8261a" (UID: "4b678a2b-2601-4267-8f78-9b6262b8261a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.511213 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities" (OuterVolumeSpecName: "utilities") pod "60d83505-44fa-47d9-b05c-def5eda227a9" (UID: "60d83505-44fa-47d9-b05c-def5eda227a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.512281 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95" (OuterVolumeSpecName: "kube-api-access-sqd95") pod "4b678a2b-2601-4267-8f78-9b6262b8261a" (UID: "4b678a2b-2601-4267-8f78-9b6262b8261a"). InnerVolumeSpecName "kube-api-access-sqd95". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.513888 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c8af8728-cae3-44d4-9f71-79c69e8c5646" (UID: "c8af8728-cae3-44d4-9f71-79c69e8c5646"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.515744 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh" (OuterVolumeSpecName: "kube-api-access-59ckh") pod "c8af8728-cae3-44d4-9f71-79c69e8c5646" (UID: "c8af8728-cae3-44d4-9f71-79c69e8c5646"). InnerVolumeSpecName "kube-api-access-59ckh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.531630 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7" (OuterVolumeSpecName: "kube-api-access-xhhm7") pod "60d83505-44fa-47d9-b05c-def5eda227a9" (UID: "60d83505-44fa-47d9-b05c-def5eda227a9"). InnerVolumeSpecName "kube-api-access-xhhm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.542662 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60d83505-44fa-47d9-b05c-def5eda227a9" (UID: "60d83505-44fa-47d9-b05c-def5eda227a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.577217 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zlnn7"] Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611378 4880 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611428 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611454 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhhm7\" (UniqueName: \"kubernetes.io/projected/60d83505-44fa-47d9-b05c-def5eda227a9-kube-api-access-xhhm7\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611464 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611473 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqd95\" (UniqueName: \"kubernetes.io/projected/4b678a2b-2601-4267-8f78-9b6262b8261a-kube-api-access-sqd95\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611482 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59ckh\" (UniqueName: \"kubernetes.io/projected/c8af8728-cae3-44d4-9f71-79c69e8c5646-kube-api-access-59ckh\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611491 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d83505-44fa-47d9-b05c-def5eda227a9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.611499 4880 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8af8728-cae3-44d4-9f71-79c69e8c5646-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.661800 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.676515 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b678a2b-2601-4267-8f78-9b6262b8261a" (UID: "4b678a2b-2601-4267-8f78-9b6262b8261a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.713085 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b678a2b-2601-4267-8f78-9b6262b8261a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.814157 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqfwj\" (UniqueName: \"kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj\") pod \"7fff77b4-8896-4c5f-8960-93aedba5de59\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.814282 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities\") pod \"7fff77b4-8896-4c5f-8960-93aedba5de59\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.814389 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content\") pod \"7fff77b4-8896-4c5f-8960-93aedba5de59\" (UID: \"7fff77b4-8896-4c5f-8960-93aedba5de59\") " Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.815258 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities" (OuterVolumeSpecName: "utilities") pod "7fff77b4-8896-4c5f-8960-93aedba5de59" (UID: "7fff77b4-8896-4c5f-8960-93aedba5de59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.818948 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj" (OuterVolumeSpecName: "kube-api-access-wqfwj") pod "7fff77b4-8896-4c5f-8960-93aedba5de59" (UID: "7fff77b4-8896-4c5f-8960-93aedba5de59"). InnerVolumeSpecName "kube-api-access-wqfwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.888858 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fff77b4-8896-4c5f-8960-93aedba5de59" (UID: "7fff77b4-8896-4c5f-8960-93aedba5de59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.915750 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqfwj\" (UniqueName: \"kubernetes.io/projected/7fff77b4-8896-4c5f-8960-93aedba5de59-kube-api-access-wqfwj\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.915806 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:49 crc kubenswrapper[4880]: I0126 16:00:49.915819 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fff77b4-8896-4c5f-8960-93aedba5de59-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.038070 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d9pw" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.038062 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d9pw" event={"ID":"7fff77b4-8896-4c5f-8960-93aedba5de59","Type":"ContainerDied","Data":"5be71e230c6932e24882571670740b203d4ca63cb3f0c58646a59c4cda9ea17c"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.038254 4880 scope.go:117] "RemoveContainer" containerID="894dac2c15b0f933800e82dffe634f9a835ead060be1a353bb26f251b2692ef0" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.042056 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r4dp4" event={"ID":"60d83505-44fa-47d9-b05c-def5eda227a9","Type":"ContainerDied","Data":"80554d42406f1e7cd460e45eaa71a861c4c8b562723a42c8dc451b8d3ba447cb"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.042163 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r4dp4" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.044384 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" event={"ID":"c376a5ed-f9eb-4592-b6b5-2e021c59d978","Type":"ContainerStarted","Data":"a5e290ce7802ce5a92e69999810eb3e7589b8ee76154b112bb0ace64d2c46397"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.044467 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" event={"ID":"c376a5ed-f9eb-4592-b6b5-2e021c59d978","Type":"ContainerStarted","Data":"4f2a92d0ada573116fd9a386fd8737668205456b8f8bb8c4cbd53e005249717b"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.045218 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.060636 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.068905 4880 scope.go:117] "RemoveContainer" containerID="74e62edc0a5e5329a7f77b265b3b3d6b05fc13b62716d944eb47b99b77e97d6d" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.073058 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9hz" event={"ID":"4b678a2b-2601-4267-8f78-9b6262b8261a","Type":"ContainerDied","Data":"f06cd8f4ea781e781cab769961c1b1286a9fddd367cc2fd2791990d0ae297321"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.073110 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9hz" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.083541 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" event={"ID":"c8af8728-cae3-44d4-9f71-79c69e8c5646","Type":"ContainerDied","Data":"adeead3640a9f2ab879c4498b469638e1e6461d35b625afba096a6f059a25c16"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.083710 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7rmh4" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.097984 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xxrk" event={"ID":"ebb3fdcf-80cb-4293-9e7d-17a44072c91e","Type":"ContainerDied","Data":"80e0f9ba0d31d0cbc5f110a12a47bcf7ea1ca2f381a15d5d2b5673884e0fa6bf"} Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.098107 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xxrk" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.105217 4880 scope.go:117] "RemoveContainer" containerID="3255312287a4bb286b16d259765e4172f95bdf38861a58674ade6c60485e494b" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.110346 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zlnn7" podStartSLOduration=2.110314734 podStartE2EDuration="2.110314734s" podCreationTimestamp="2026-01-26 16:00:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:00:50.092998804 +0000 UTC m=+375.578727511" watchObservedRunningTime="2026-01-26 16:00:50.110314734 +0000 UTC m=+375.596043441" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.116400 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.121664 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5d9pw"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.141534 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.145717 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r4dp4"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.161127 4880 scope.go:117] "RemoveContainer" containerID="2147c5413f25743ec2170f5006969eb3d9047ad21d57a3b4d7f87ee419bbf6a8" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.187655 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.188701 4880 scope.go:117] "RemoveContainer" containerID="25e8984948d1d66f4e2dc7f79754d5b3cda3483855571b1a90af19a554227884" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.192069 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qr9hz"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.209785 4880 scope.go:117] "RemoveContainer" containerID="599a6329ec7869ff418b2aae242e419ecde6e7a5313fe73db51f41e7943cdd49" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.209847 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.230043 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" path="/var/lib/kubelet/pods/4b678a2b-2601-4267-8f78-9b6262b8261a/volumes" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.231830 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" path="/var/lib/kubelet/pods/60d83505-44fa-47d9-b05c-def5eda227a9/volumes" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.233998 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" path="/var/lib/kubelet/pods/7fff77b4-8896-4c5f-8960-93aedba5de59/volumes" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.236014 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7rmh4"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.236283 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.236902 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4xxrk"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.246900 4880 scope.go:117] "RemoveContainer" containerID="30c7c6d17d43b2796d289522c26623a2bf6197f2d1b56093a53c0a73db9c52af" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.262623 4880 scope.go:117] "RemoveContainer" containerID="73cd4f146134d887516a8bfdcfd1026653df0964a1333bf770acc797996be2e1" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.287065 4880 scope.go:117] "RemoveContainer" containerID="08daebfe2d6811c42521510ead7d6c2d2d043b01ff1bdd8de86c5a6dd707dc76" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.315457 4880 scope.go:117] "RemoveContainer" containerID="bb2d56aac038c8b94316087d7e36c62df80a79abf07acb027f55b010328daf6e" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.330932 4880 scope.go:117] "RemoveContainer" containerID="e9ef9061b9c1b44f6004f97bf27b70428342a7d264d19f1e1005a51c311bd149" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.345261 4880 scope.go:117] "RemoveContainer" containerID="95f8ea210ee99020c0c6911532f871c1875e7a2b4f705fddd0c034a7477eb92f" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.366937 4880 scope.go:117] "RemoveContainer" containerID="3f580137fad1bdca344dcf575ca400505815ad0151674bcdac16af4f8daac7c3" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.792079 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.793651 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.793780 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.793877 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.793993 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.794081 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.794165 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.794282 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.794370 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.794499 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.794620 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.794725 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.794812 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.794942 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.795034 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.795117 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.795219 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.795305 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.795382 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="extract-utilities" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.795490 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.795598 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.795684 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.795801 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="extract-content" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.795942 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796064 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: E0126 16:00:50.796160 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796271 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796524 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b678a2b-2601-4267-8f78-9b6262b8261a" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796662 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" containerName="marketplace-operator" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796775 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d83505-44fa-47d9-b05c-def5eda227a9" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796901 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.796998 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fff77b4-8896-4c5f-8960-93aedba5de59" containerName="registry-server" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.798334 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.802612 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.806621 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.951579 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.953542 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtsrf\" (UniqueName: \"kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.953982 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.992333 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mkk48"] Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.993769 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:50 crc kubenswrapper[4880]: I0126 16:00:50.998805 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.004100 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkk48"] Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055021 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055082 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-utilities\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055187 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkskf\" (UniqueName: \"kubernetes.io/projected/9ec96499-4381-4628-9c64-1a8de843bf6d-kube-api-access-kkskf\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055258 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055312 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-catalog-content\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055373 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtsrf\" (UniqueName: \"kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.055636 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.056017 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.079570 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtsrf\" (UniqueName: \"kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf\") pod \"redhat-marketplace-t7qhx\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.128848 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.156922 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-catalog-content\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.157054 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-utilities\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.157082 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkskf\" (UniqueName: \"kubernetes.io/projected/9ec96499-4381-4628-9c64-1a8de843bf6d-kube-api-access-kkskf\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.157784 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-catalog-content\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.158324 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ec96499-4381-4628-9c64-1a8de843bf6d-utilities\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.178952 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkskf\" (UniqueName: \"kubernetes.io/projected/9ec96499-4381-4628-9c64-1a8de843bf6d-kube-api-access-kkskf\") pod \"redhat-operators-mkk48\" (UID: \"9ec96499-4381-4628-9c64-1a8de843bf6d\") " pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.319357 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.560255 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:00:51 crc kubenswrapper[4880]: W0126 16:00:51.568847 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c79ba18_65a0_4ef8_87e9_1db08451a4e6.slice/crio-217e35f34cf348e472b5289f0377bbb10769a8ea301539d27204f184e0223a2c WatchSource:0}: Error finding container 217e35f34cf348e472b5289f0377bbb10769a8ea301539d27204f184e0223a2c: Status 404 returned error can't find the container with id 217e35f34cf348e472b5289f0377bbb10769a8ea301539d27204f184e0223a2c Jan 26 16:00:51 crc kubenswrapper[4880]: I0126 16:00:51.742248 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkk48"] Jan 26 16:00:51 crc kubenswrapper[4880]: W0126 16:00:51.747398 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ec96499_4381_4628_9c64_1a8de843bf6d.slice/crio-d7b0023b972311665b2dba5e31927b011302aea3bb34453db98a934401ee6a4b WatchSource:0}: Error finding container d7b0023b972311665b2dba5e31927b011302aea3bb34453db98a934401ee6a4b: Status 404 returned error can't find the container with id d7b0023b972311665b2dba5e31927b011302aea3bb34453db98a934401ee6a4b Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.114890 4880 generic.go:334] "Generic (PLEG): container finished" podID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerID="298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b" exitCode=0 Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.114993 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerDied","Data":"298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b"} Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.115255 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerStarted","Data":"217e35f34cf348e472b5289f0377bbb10769a8ea301539d27204f184e0223a2c"} Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.119045 4880 generic.go:334] "Generic (PLEG): container finished" podID="9ec96499-4381-4628-9c64-1a8de843bf6d" containerID="07cb17ae67b67ace6b76e81a97358bfe3d14dd2d99ecd60d550e15d90a8df124" exitCode=0 Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.119130 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkk48" event={"ID":"9ec96499-4381-4628-9c64-1a8de843bf6d","Type":"ContainerDied","Data":"07cb17ae67b67ace6b76e81a97358bfe3d14dd2d99ecd60d550e15d90a8df124"} Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.119216 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkk48" event={"ID":"9ec96499-4381-4628-9c64-1a8de843bf6d","Type":"ContainerStarted","Data":"d7b0023b972311665b2dba5e31927b011302aea3bb34453db98a934401ee6a4b"} Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.222375 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8af8728-cae3-44d4-9f71-79c69e8c5646" path="/var/lib/kubelet/pods/c8af8728-cae3-44d4-9f71-79c69e8c5646/volumes" Jan 26 16:00:52 crc kubenswrapper[4880]: I0126 16:00:52.223083 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebb3fdcf-80cb-4293-9e7d-17a44072c91e" path="/var/lib/kubelet/pods/ebb3fdcf-80cb-4293-9e7d-17a44072c91e/volumes" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.199716 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p7xk2"] Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.201399 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.204586 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.208555 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7xk2"] Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.300078 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-catalog-content\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.300131 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lglf2\" (UniqueName: \"kubernetes.io/projected/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-kube-api-access-lglf2\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.300328 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-utilities\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.475140 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-utilities\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.475218 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-catalog-content\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.475253 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lglf2\" (UniqueName: \"kubernetes.io/projected/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-kube-api-access-lglf2\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.476164 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-utilities\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.477243 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-catalog-content\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.495792 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zksbn"] Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.497083 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.499839 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.514883 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zksbn"] Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.518230 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lglf2\" (UniqueName: \"kubernetes.io/projected/6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac-kube-api-access-lglf2\") pod \"certified-operators-p7xk2\" (UID: \"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac\") " pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.537833 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.722177 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-utilities\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.722927 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzm2\" (UniqueName: \"kubernetes.io/projected/4176b049-00e6-4f38-8045-4d84d2926272-kube-api-access-hpzm2\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.722975 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-catalog-content\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.824946 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-utilities\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.825012 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzm2\" (UniqueName: \"kubernetes.io/projected/4176b049-00e6-4f38-8045-4d84d2926272-kube-api-access-hpzm2\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.825033 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-catalog-content\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.825600 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-utilities\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.825663 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4176b049-00e6-4f38-8045-4d84d2926272-catalog-content\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.851749 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzm2\" (UniqueName: \"kubernetes.io/projected/4176b049-00e6-4f38-8045-4d84d2926272-kube-api-access-hpzm2\") pod \"community-operators-zksbn\" (UID: \"4176b049-00e6-4f38-8045-4d84d2926272\") " pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:53 crc kubenswrapper[4880]: I0126 16:00:53.957040 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.071340 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p7xk2"] Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.139359 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7xk2" event={"ID":"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac","Type":"ContainerStarted","Data":"3bf078c2b0d9d5ab6fa30afe09b08ef59c660b03ea1cef2355a7df2995940303"} Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.142760 4880 generic.go:334] "Generic (PLEG): container finished" podID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerID="e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f" exitCode=0 Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.142920 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerDied","Data":"e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f"} Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.239226 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkk48" event={"ID":"9ec96499-4381-4628-9c64-1a8de843bf6d","Type":"ContainerStarted","Data":"175d3301e9e669193595202f7371e779b34181ee325f3a6ab372e604e2382fa8"} Jan 26 16:00:54 crc kubenswrapper[4880]: I0126 16:00:54.247903 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zksbn"] Jan 26 16:00:54 crc kubenswrapper[4880]: W0126 16:00:54.264582 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4176b049_00e6_4f38_8045_4d84d2926272.slice/crio-e58ef13d8e22c57ac0a1d3e26c42123fceedd4709efbcdcb058c66dd017c3fa1 WatchSource:0}: Error finding container e58ef13d8e22c57ac0a1d3e26c42123fceedd4709efbcdcb058c66dd017c3fa1: Status 404 returned error can't find the container with id e58ef13d8e22c57ac0a1d3e26c42123fceedd4709efbcdcb058c66dd017c3fa1 Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.243045 4880 generic.go:334] "Generic (PLEG): container finished" podID="6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac" containerID="438261955d46c50eaf09df4db75c28bfe54db6e56decc0d6cbef2f6c26724a88" exitCode=0 Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.243137 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7xk2" event={"ID":"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac","Type":"ContainerDied","Data":"438261955d46c50eaf09df4db75c28bfe54db6e56decc0d6cbef2f6c26724a88"} Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.245464 4880 generic.go:334] "Generic (PLEG): container finished" podID="4176b049-00e6-4f38-8045-4d84d2926272" containerID="d02ae26627d97ca2fa915fe99a205e07b1d253b40f74c23d499196709e701c64" exitCode=0 Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.245534 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zksbn" event={"ID":"4176b049-00e6-4f38-8045-4d84d2926272","Type":"ContainerDied","Data":"d02ae26627d97ca2fa915fe99a205e07b1d253b40f74c23d499196709e701c64"} Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.245678 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zksbn" event={"ID":"4176b049-00e6-4f38-8045-4d84d2926272","Type":"ContainerStarted","Data":"e58ef13d8e22c57ac0a1d3e26c42123fceedd4709efbcdcb058c66dd017c3fa1"} Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.248688 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerStarted","Data":"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805"} Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.252943 4880 generic.go:334] "Generic (PLEG): container finished" podID="9ec96499-4381-4628-9c64-1a8de843bf6d" containerID="175d3301e9e669193595202f7371e779b34181ee325f3a6ab372e604e2382fa8" exitCode=0 Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.252998 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkk48" event={"ID":"9ec96499-4381-4628-9c64-1a8de843bf6d","Type":"ContainerDied","Data":"175d3301e9e669193595202f7371e779b34181ee325f3a6ab372e604e2382fa8"} Jan 26 16:00:55 crc kubenswrapper[4880]: I0126 16:00:55.297969 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7qhx" podStartSLOduration=2.5327779660000003 podStartE2EDuration="5.297906529s" podCreationTimestamp="2026-01-26 16:00:50 +0000 UTC" firstStartedPulling="2026-01-26 16:00:52.116402655 +0000 UTC m=+377.602131362" lastFinishedPulling="2026-01-26 16:00:54.881531218 +0000 UTC m=+380.367259925" observedRunningTime="2026-01-26 16:00:55.29350201 +0000 UTC m=+380.779230727" watchObservedRunningTime="2026-01-26 16:00:55.297906529 +0000 UTC m=+380.783635236" Jan 26 16:00:56 crc kubenswrapper[4880]: I0126 16:00:56.262759 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkk48" event={"ID":"9ec96499-4381-4628-9c64-1a8de843bf6d","Type":"ContainerStarted","Data":"b8a2942e4ce4b6cda3167524bfe5126e85f2a77fe18d8defcaee9538e4887b4f"} Jan 26 16:00:56 crc kubenswrapper[4880]: I0126 16:00:56.288929 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mkk48" podStartSLOduration=2.538543934 podStartE2EDuration="6.288909053s" podCreationTimestamp="2026-01-26 16:00:50 +0000 UTC" firstStartedPulling="2026-01-26 16:00:52.120577118 +0000 UTC m=+377.606305825" lastFinishedPulling="2026-01-26 16:00:55.870942237 +0000 UTC m=+381.356670944" observedRunningTime="2026-01-26 16:00:56.286942959 +0000 UTC m=+381.772671666" watchObservedRunningTime="2026-01-26 16:00:56.288909053 +0000 UTC m=+381.774637760" Jan 26 16:00:57 crc kubenswrapper[4880]: I0126 16:00:57.292504 4880 generic.go:334] "Generic (PLEG): container finished" podID="6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac" containerID="dcdc552121d5eb06c41f6e878e0903fbf461f7ea0cc8b564fe7ac351733ebf90" exitCode=0 Jan 26 16:00:57 crc kubenswrapper[4880]: I0126 16:00:57.292700 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7xk2" event={"ID":"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac","Type":"ContainerDied","Data":"dcdc552121d5eb06c41f6e878e0903fbf461f7ea0cc8b564fe7ac351733ebf90"} Jan 26 16:00:57 crc kubenswrapper[4880]: I0126 16:00:57.296743 4880 generic.go:334] "Generic (PLEG): container finished" podID="4176b049-00e6-4f38-8045-4d84d2926272" containerID="e74bf20ea4251e453c228062543c495edb009ba00349b82dc57327ba90196aa0" exitCode=0 Jan 26 16:00:57 crc kubenswrapper[4880]: I0126 16:00:57.296806 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zksbn" event={"ID":"4176b049-00e6-4f38-8045-4d84d2926272","Type":"ContainerDied","Data":"e74bf20ea4251e453c228062543c495edb009ba00349b82dc57327ba90196aa0"} Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.311605 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zksbn" event={"ID":"4176b049-00e6-4f38-8045-4d84d2926272","Type":"ContainerStarted","Data":"e26d9fd74bb82061fa0fb7ce02209250b962029c264ada1bdc801f36d4c615f4"} Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.317907 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p7xk2" event={"ID":"6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac","Type":"ContainerStarted","Data":"b3e0d61ab13aa8d49208b022028b798132e6cba2fb61b5e002333bfd3ff6f70e"} Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.351036 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zksbn" podStartSLOduration=2.67815827 podStartE2EDuration="5.351013225s" podCreationTimestamp="2026-01-26 16:00:53 +0000 UTC" firstStartedPulling="2026-01-26 16:00:55.24679601 +0000 UTC m=+380.732524717" lastFinishedPulling="2026-01-26 16:00:57.919650965 +0000 UTC m=+383.405379672" observedRunningTime="2026-01-26 16:00:58.346870263 +0000 UTC m=+383.832598970" watchObservedRunningTime="2026-01-26 16:00:58.351013225 +0000 UTC m=+383.836741932" Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.367906 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p7xk2" podStartSLOduration=2.621279685 podStartE2EDuration="5.367888714s" podCreationTimestamp="2026-01-26 16:00:53 +0000 UTC" firstStartedPulling="2026-01-26 16:00:55.244902329 +0000 UTC m=+380.730631036" lastFinishedPulling="2026-01-26 16:00:57.991511358 +0000 UTC m=+383.477240065" observedRunningTime="2026-01-26 16:00:58.366892746 +0000 UTC m=+383.852621453" watchObservedRunningTime="2026-01-26 16:00:58.367888714 +0000 UTC m=+383.853617421" Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.621829 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-p9gv9" Jan 26 16:00:58 crc kubenswrapper[4880]: I0126 16:00:58.685023 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.129901 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.130022 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.184036 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.320702 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.320778 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:01:01 crc kubenswrapper[4880]: I0126 16:01:01.375214 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:01:02 crc kubenswrapper[4880]: I0126 16:01:02.389238 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mkk48" podUID="9ec96499-4381-4628-9c64-1a8de843bf6d" containerName="registry-server" probeResult="failure" output=< Jan 26 16:01:02 crc kubenswrapper[4880]: timeout: failed to connect service ":50051" within 1s Jan 26 16:01:02 crc kubenswrapper[4880]: > Jan 26 16:01:03 crc kubenswrapper[4880]: I0126 16:01:03.538132 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:01:03 crc kubenswrapper[4880]: I0126 16:01:03.538199 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:01:03 crc kubenswrapper[4880]: I0126 16:01:03.603175 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:01:03 crc kubenswrapper[4880]: I0126 16:01:03.958287 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:01:03 crc kubenswrapper[4880]: I0126 16:01:03.958352 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:01:04 crc kubenswrapper[4880]: I0126 16:01:04.005043 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:01:04 crc kubenswrapper[4880]: I0126 16:01:04.488399 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zksbn" Jan 26 16:01:04 crc kubenswrapper[4880]: I0126 16:01:04.497155 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p7xk2" Jan 26 16:01:10 crc kubenswrapper[4880]: I0126 16:01:10.700325 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:01:10 crc kubenswrapper[4880]: I0126 16:01:10.700969 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:01:11 crc kubenswrapper[4880]: I0126 16:01:11.367090 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:01:11 crc kubenswrapper[4880]: I0126 16:01:11.408395 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mkk48" Jan 26 16:01:12 crc kubenswrapper[4880]: I0126 16:01:12.386797 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:01:12 crc kubenswrapper[4880]: I0126 16:01:12.387094 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerName="controller-manager" containerID="cri-o://e4e636a9ad1440b1fd928900036986c6d7ce6d37863a7708776d980ac9874f56" gracePeriod=30 Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.308060 4880 patch_prober.go:28] interesting pod/controller-manager-789d9f7b94-zk4ts container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.308158 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.465650 4880 generic.go:334] "Generic (PLEG): container finished" podID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerID="e4e636a9ad1440b1fd928900036986c6d7ce6d37863a7708776d980ac9874f56" exitCode=0 Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.465860 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" event={"ID":"a70cd4a9-e2ee-456d-a032-9fb0607dc91a","Type":"ContainerDied","Data":"e4e636a9ad1440b1fd928900036986c6d7ce6d37863a7708776d980ac9874f56"} Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.648827 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.683568 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56448fb7db-v6gf5"] Jan 26 16:01:14 crc kubenswrapper[4880]: E0126 16:01:14.683906 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerName="controller-manager" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.683925 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerName="controller-manager" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.684157 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" containerName="controller-manager" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.684738 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.695231 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles\") pod \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.695317 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert\") pod \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.695358 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca\") pod \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.695422 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config\") pod \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.695511 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swxph\" (UniqueName: \"kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph\") pod \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\" (UID: \"a70cd4a9-e2ee-456d-a032-9fb0607dc91a\") " Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.698618 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config" (OuterVolumeSpecName: "config") pod "a70cd4a9-e2ee-456d-a032-9fb0607dc91a" (UID: "a70cd4a9-e2ee-456d-a032-9fb0607dc91a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.698672 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca" (OuterVolumeSpecName: "client-ca") pod "a70cd4a9-e2ee-456d-a032-9fb0607dc91a" (UID: "a70cd4a9-e2ee-456d-a032-9fb0607dc91a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.699116 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a70cd4a9-e2ee-456d-a032-9fb0607dc91a" (UID: "a70cd4a9-e2ee-456d-a032-9fb0607dc91a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.700294 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56448fb7db-v6gf5"] Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.704031 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph" (OuterVolumeSpecName: "kube-api-access-swxph") pod "a70cd4a9-e2ee-456d-a032-9fb0607dc91a" (UID: "a70cd4a9-e2ee-456d-a032-9fb0607dc91a"). InnerVolumeSpecName "kube-api-access-swxph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.705302 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a70cd4a9-e2ee-456d-a032-9fb0607dc91a" (UID: "a70cd4a9-e2ee-456d-a032-9fb0607dc91a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797190 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/032fa725-90a7-4d50-8836-f9e80c58282e-serving-cert\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797255 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-proxy-ca-bundles\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797298 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-client-ca\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797323 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g8hr\" (UniqueName: \"kubernetes.io/projected/032fa725-90a7-4d50-8836-f9e80c58282e-kube-api-access-4g8hr\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797499 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-config\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797645 4880 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797662 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swxph\" (UniqueName: \"kubernetes.io/projected/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-kube-api-access-swxph\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797674 4880 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797684 4880 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.797693 4880 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a70cd4a9-e2ee-456d-a032-9fb0607dc91a-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.898518 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-proxy-ca-bundles\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.898599 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-client-ca\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.898635 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g8hr\" (UniqueName: \"kubernetes.io/projected/032fa725-90a7-4d50-8836-f9e80c58282e-kube-api-access-4g8hr\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.898685 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-config\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.898746 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/032fa725-90a7-4d50-8836-f9e80c58282e-serving-cert\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.900628 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-proxy-ca-bundles\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.900857 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-client-ca\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.902048 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032fa725-90a7-4d50-8836-f9e80c58282e-config\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.904675 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/032fa725-90a7-4d50-8836-f9e80c58282e-serving-cert\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:14 crc kubenswrapper[4880]: I0126 16:01:14.919943 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g8hr\" (UniqueName: \"kubernetes.io/projected/032fa725-90a7-4d50-8836-f9e80c58282e-kube-api-access-4g8hr\") pod \"controller-manager-56448fb7db-v6gf5\" (UID: \"032fa725-90a7-4d50-8836-f9e80c58282e\") " pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.034314 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.473313 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" event={"ID":"a70cd4a9-e2ee-456d-a032-9fb0607dc91a","Type":"ContainerDied","Data":"e50286eead330331a71ef1742053652f57f48204757d3dfb7201528d0e82ff95"} Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.473403 4880 scope.go:117] "RemoveContainer" containerID="e4e636a9ad1440b1fd928900036986c6d7ce6d37863a7708776d980ac9874f56" Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.473463 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d9f7b94-zk4ts" Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.650487 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.662212 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-789d9f7b94-zk4ts"] Jan 26 16:01:15 crc kubenswrapper[4880]: I0126 16:01:15.689527 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56448fb7db-v6gf5"] Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.221492 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70cd4a9-e2ee-456d-a032-9fb0607dc91a" path="/var/lib/kubelet/pods/a70cd4a9-e2ee-456d-a032-9fb0607dc91a/volumes" Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.482681 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" event={"ID":"032fa725-90a7-4d50-8836-f9e80c58282e","Type":"ContainerStarted","Data":"e0b2131c6b59447f037a4b409780516e74c9d18784fb11032208fe45bacc906f"} Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.482734 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" event={"ID":"032fa725-90a7-4d50-8836-f9e80c58282e","Type":"ContainerStarted","Data":"df081197393b599350cee6a5bab4134e9631c5d8e8deb912f1457b78c6d0bce9"} Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.485633 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.514213 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" podStartSLOduration=4.514176849 podStartE2EDuration="4.514176849s" podCreationTimestamp="2026-01-26 16:01:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:01:16.513217874 +0000 UTC m=+401.998946581" watchObservedRunningTime="2026-01-26 16:01:16.514176849 +0000 UTC m=+401.999905566" Jan 26 16:01:16 crc kubenswrapper[4880]: I0126 16:01:16.521951 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56448fb7db-v6gf5" Jan 26 16:01:23 crc kubenswrapper[4880]: I0126 16:01:23.810984 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" podUID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" containerName="registry" containerID="cri-o://38eca27fc527afe3bb1284d98916ecd88c77aa5969c29ead80b761ff6ee37bb9" gracePeriod=30 Jan 26 16:01:24 crc kubenswrapper[4880]: I0126 16:01:24.565829 4880 generic.go:334] "Generic (PLEG): container finished" podID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" containerID="38eca27fc527afe3bb1284d98916ecd88c77aa5969c29ead80b761ff6ee37bb9" exitCode=0 Jan 26 16:01:24 crc kubenswrapper[4880]: I0126 16:01:24.565953 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" event={"ID":"5e2725c4-874b-4e90-a0c1-ade113a0b7e2","Type":"ContainerDied","Data":"38eca27fc527afe3bb1284d98916ecd88c77aa5969c29ead80b761ff6ee37bb9"} Jan 26 16:01:24 crc kubenswrapper[4880]: I0126 16:01:24.930094 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085127 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wcz6\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085218 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085257 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085304 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085332 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085649 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085673 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.085733 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token\") pod \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\" (UID: \"5e2725c4-874b-4e90-a0c1-ade113a0b7e2\") " Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.087282 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.087408 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.092956 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.092941 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.093683 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6" (OuterVolumeSpecName: "kube-api-access-6wcz6") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "kube-api-access-6wcz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.093885 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.098455 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.107286 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5e2725c4-874b-4e90-a0c1-ade113a0b7e2" (UID: "5e2725c4-874b-4e90-a0c1-ade113a0b7e2"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187628 4880 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187740 4880 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187751 4880 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187766 4880 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187775 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wcz6\" (UniqueName: \"kubernetes.io/projected/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-kube-api-access-6wcz6\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187784 4880 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.187821 4880 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5e2725c4-874b-4e90-a0c1-ade113a0b7e2-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.574893 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" event={"ID":"5e2725c4-874b-4e90-a0c1-ade113a0b7e2","Type":"ContainerDied","Data":"89e2a3313aa26f2668d7aef6cb58afbb222f543405db70d5b8e27ae349457469"} Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.574968 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zc2dp" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.575003 4880 scope.go:117] "RemoveContainer" containerID="38eca27fc527afe3bb1284d98916ecd88c77aa5969c29ead80b761ff6ee37bb9" Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.609651 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 16:01:25 crc kubenswrapper[4880]: I0126 16:01:25.614272 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zc2dp"] Jan 26 16:01:26 crc kubenswrapper[4880]: I0126 16:01:26.223303 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" path="/var/lib/kubelet/pods/5e2725c4-874b-4e90-a0c1-ade113a0b7e2/volumes" Jan 26 16:01:40 crc kubenswrapper[4880]: I0126 16:01:40.699857 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:01:40 crc kubenswrapper[4880]: I0126 16:01:40.700591 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:01:40 crc kubenswrapper[4880]: I0126 16:01:40.700656 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:01:40 crc kubenswrapper[4880]: I0126 16:01:40.701720 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:01:40 crc kubenswrapper[4880]: I0126 16:01:40.701792 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a" gracePeriod=600 Jan 26 16:01:41 crc kubenswrapper[4880]: I0126 16:01:41.735719 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a" exitCode=0 Jan 26 16:01:41 crc kubenswrapper[4880]: I0126 16:01:41.735819 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a"} Jan 26 16:01:41 crc kubenswrapper[4880]: I0126 16:01:41.736077 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585"} Jan 26 16:01:41 crc kubenswrapper[4880]: I0126 16:01:41.736122 4880 scope.go:117] "RemoveContainer" containerID="5f536a2ab9ffd53a879c5ab0528faafb622a1c1dab5045acbd7a59b11abbb1ed" Jan 26 16:03:40 crc kubenswrapper[4880]: I0126 16:03:40.700357 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:03:40 crc kubenswrapper[4880]: I0126 16:03:40.701023 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:04:10 crc kubenswrapper[4880]: I0126 16:04:10.699783 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:04:10 crc kubenswrapper[4880]: I0126 16:04:10.700509 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:04:40 crc kubenswrapper[4880]: I0126 16:04:40.699925 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:04:40 crc kubenswrapper[4880]: I0126 16:04:40.701749 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:04:40 crc kubenswrapper[4880]: I0126 16:04:40.701917 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:04:40 crc kubenswrapper[4880]: I0126 16:04:40.702585 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:04:40 crc kubenswrapper[4880]: I0126 16:04:40.702744 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585" gracePeriod=600 Jan 26 16:04:41 crc kubenswrapper[4880]: I0126 16:04:41.064020 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585" exitCode=0 Jan 26 16:04:41 crc kubenswrapper[4880]: I0126 16:04:41.064117 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585"} Jan 26 16:04:41 crc kubenswrapper[4880]: I0126 16:04:41.064378 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da"} Jan 26 16:04:41 crc kubenswrapper[4880]: I0126 16:04:41.064425 4880 scope.go:117] "RemoveContainer" containerID="ee33004dd7545d28c32edb1f9e46bc586c34f01596e2d8b57dd0d73a8edcee0a" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.065789 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2p7jp"] Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068598 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068598 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="nbdb" containerID="cri-o://e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068686 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="sbdb" containerID="cri-o://f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068660 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="northd" containerID="cri-o://c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068919 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-node" containerID="cri-o://0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.068950 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-acl-logging" containerID="cri-o://c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.079036 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-controller" containerID="cri-o://5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.131073 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" containerID="cri-o://662f18544e42fc1d98d1b06fe7e06c28caf59e4a88a8701f25a28195713bb3d9" gracePeriod=30 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.438959 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/1.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.439507 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/0.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.439572 4880 generic.go:334] "Generic (PLEG): container finished" podID="ddcd016a-fdad-4580-adaf-30451ac85f0e" containerID="7963c3cc01b50b41810def1f17a183f3f03df1ec0d64b079bda8841b0e6f900c" exitCode=2 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.439657 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerDied","Data":"7963c3cc01b50b41810def1f17a183f3f03df1ec0d64b079bda8841b0e6f900c"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.439801 4880 scope.go:117] "RemoveContainer" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.440417 4880 scope.go:117] "RemoveContainer" containerID="7963c3cc01b50b41810def1f17a183f3f03df1ec0d64b079bda8841b0e6f900c" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.440686 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-h6f8z_openshift-multus(ddcd016a-fdad-4580-adaf-30451ac85f0e)\"" pod="openshift-multus/multus-h6f8z" podUID="ddcd016a-fdad-4580-adaf-30451ac85f0e" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.448035 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovnkube-controller/2.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.451288 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-acl-logging/0.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.451807 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-controller/0.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452171 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="662f18544e42fc1d98d1b06fe7e06c28caf59e4a88a8701f25a28195713bb3d9" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452198 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452205 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452214 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452222 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452231 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b" exitCode=0 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452241 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f" exitCode=143 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452249 4880 generic.go:334] "Generic (PLEG): container finished" podID="c377ad66-a182-4cbf-9b29-253f3e547118" containerID="5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203" exitCode=143 Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452255 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"662f18544e42fc1d98d1b06fe7e06c28caf59e4a88a8701f25a28195713bb3d9"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452297 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452311 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452321 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452356 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452365 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452376 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.452387 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203"} Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.636476 4880 scope.go:117] "RemoveContainer" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.659878 4880 scope.go:117] "RemoveContainer" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.668784 4880 scope.go:117] "RemoveContainer" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.669416 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\": container with ID starting with b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232 not found: ID does not exist" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.669498 4880 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\": rpc error: code = NotFound desc = could not find container \"b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232\": container with ID starting with b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232 not found: ID does not exist" containerID="b316e47a6940219c3645e38a0de529b2260ad94141825a9380ad853d3fd5f232" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.672221 4880 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_ovnkube-controller_ovnkube-node-2p7jp_openshift-ovn-kubernetes_c377ad66-a182-4cbf-9b29-253f3e547118_2 in pod sandbox 7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf from index: no such id: '5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a'" containerID="5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.672286 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a"} err="rpc error: code = Unknown desc = failed to delete container k8s_ovnkube-controller_ovnkube-node-2p7jp_openshift-ovn-kubernetes_c377ad66-a182-4cbf-9b29-253f3e547118_2 in pod sandbox 7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf from index: no such id: '5301232b6b22fc8d0f610aabd9fea769a4472fd4dd321e0ff578ddc3f30c146a'" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.834405 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-acl-logging/0.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.834917 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-controller/0.log" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.835380 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.861215 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.861885 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.861980 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862023 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862059 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862105 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862133 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862174 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862210 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862404 4880 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862557 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862595 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862680 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862689 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket" (OuterVolumeSpecName: "log-socket") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.862749 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log" (OuterVolumeSpecName: "node-log") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.874804 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.882133 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.900853 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c6qlc"] Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901162 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-node" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901185 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-node" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901200 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901211 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901221 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-acl-logging" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901226 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-acl-logging" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901234 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="northd" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901240 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="northd" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901252 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901258 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901264 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="nbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901270 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="nbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901277 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" containerName="registry" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901283 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" containerName="registry" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901291 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901299 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901309 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901315 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901321 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901327 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901337 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="sbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901344 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="sbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901353 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kubecfg-setup" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901359 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kubecfg-setup" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901506 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="nbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901522 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901530 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-node" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901540 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901551 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovn-acl-logging" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901558 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="northd" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901564 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="sbdb" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901570 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901577 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2725c4-874b-4e90-a0c1-ade113a0b7e2" containerName="registry" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901584 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901592 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: E0126 16:05:37.901686 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901694 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.901804 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" containerName="ovnkube-controller" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.904101 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963505 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963613 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963639 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963673 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963701 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85xjf\" (UniqueName: \"kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963724 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963708 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963780 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash" (OuterVolumeSpecName: "host-slash") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963738 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963831 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963864 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963869 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.963902 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964001 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964027 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964049 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964070 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch\") pod \"c377ad66-a182-4cbf-9b29-253f3e547118\" (UID: \"c377ad66-a182-4cbf-9b29-253f3e547118\") " Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964238 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-bin\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964264 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964279 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-ovn\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964300 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964295 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964323 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-netns\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964346 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-etc-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964379 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-kubelet\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964418 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964463 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964480 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964466 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964592 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-netd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964794 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964809 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-script-lib\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964895 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-log-socket\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964929 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-systemd-units\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.964968 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965017 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d85f3c59-f9e9-4a12-8039-f9396436a803-ovn-node-metrics-cert\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965065 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-env-overrides\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965091 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-slash\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965139 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-config\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965186 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-var-lib-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965225 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-node-log\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965272 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-systemd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965346 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jf82\" (UniqueName: \"kubernetes.io/projected/d85f3c59-f9e9-4a12-8039-f9396436a803-kube-api-access-2jf82\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965464 4880 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965487 4880 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965500 4880 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965518 4880 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965529 4880 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-slash\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965540 4880 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965551 4880 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-log-socket\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965563 4880 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965577 4880 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965609 4880 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965623 4880 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965643 4880 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-node-log\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965654 4880 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c377ad66-a182-4cbf-9b29-253f3e547118-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965666 4880 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c377ad66-a182-4cbf-9b29-253f3e547118-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965676 4880 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965687 4880 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965699 4880 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.965711 4880 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c377ad66-a182-4cbf-9b29-253f3e547118-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:37 crc kubenswrapper[4880]: I0126 16:05:37.968181 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf" (OuterVolumeSpecName: "kube-api-access-85xjf") pod "c377ad66-a182-4cbf-9b29-253f3e547118" (UID: "c377ad66-a182-4cbf-9b29-253f3e547118"). InnerVolumeSpecName "kube-api-access-85xjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067128 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-bin\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067194 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-ovn\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067218 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067426 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-netns\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067463 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-etc-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067477 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-kubelet\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067580 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-bin\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067604 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-etc-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067669 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-netns\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067651 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067750 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067771 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-kubelet\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067676 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-ovn\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067529 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067945 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-netd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.067988 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-script-lib\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068014 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-log-socket\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068055 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-systemd-units\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068119 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068123 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-log-socket\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068158 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d85f3c59-f9e9-4a12-8039-f9396436a803-ovn-node-metrics-cert\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068178 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-systemd-units\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068157 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-cni-netd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068203 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-run-ovn-kubernetes\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068235 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-env-overrides\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068270 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-slash\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068309 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-config\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068342 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-var-lib-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068387 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-node-log\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068391 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-host-slash\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068464 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-node-log\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068482 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-systemd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068519 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-var-lib-openvswitch\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068499 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d85f3c59-f9e9-4a12-8039-f9396436a803-run-systemd\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068574 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jf82\" (UniqueName: \"kubernetes.io/projected/d85f3c59-f9e9-4a12-8039-f9396436a803-kube-api-access-2jf82\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.068662 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85xjf\" (UniqueName: \"kubernetes.io/projected/c377ad66-a182-4cbf-9b29-253f3e547118-kube-api-access-85xjf\") on node \"crc\" DevicePath \"\"" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.069146 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-env-overrides\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.069386 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-config\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.070131 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d85f3c59-f9e9-4a12-8039-f9396436a803-ovnkube-script-lib\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.072775 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d85f3c59-f9e9-4a12-8039-f9396436a803-ovn-node-metrics-cert\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.087171 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jf82\" (UniqueName: \"kubernetes.io/projected/d85f3c59-f9e9-4a12-8039-f9396436a803-kube-api-access-2jf82\") pod \"ovnkube-node-c6qlc\" (UID: \"d85f3c59-f9e9-4a12-8039-f9396436a803\") " pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.220017 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.460115 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/1.log" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.461912 4880 generic.go:334] "Generic (PLEG): container finished" podID="d85f3c59-f9e9-4a12-8039-f9396436a803" containerID="578a5e78d5b3508e6864338fc2dd96541c06ec016246ae51876c94fe0616a992" exitCode=0 Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.461975 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerDied","Data":"578a5e78d5b3508e6864338fc2dd96541c06ec016246ae51876c94fe0616a992"} Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.462048 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"2957da2e1e19120ec0945b98d6b9faeb4a155e3d9837d336b874bced220d4b18"} Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.470368 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-acl-logging/0.log" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.471013 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2p7jp_c377ad66-a182-4cbf-9b29-253f3e547118/ovn-controller/0.log" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.471516 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" event={"ID":"c377ad66-a182-4cbf-9b29-253f3e547118","Type":"ContainerDied","Data":"7d9e9e43afa8f5654ded97fda1e2c919ec5239b1e6a64d61be49ca0518eb3daf"} Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.471605 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2p7jp" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.471614 4880 scope.go:117] "RemoveContainer" containerID="662f18544e42fc1d98d1b06fe7e06c28caf59e4a88a8701f25a28195713bb3d9" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.502973 4880 scope.go:117] "RemoveContainer" containerID="f6ce2eb083cde9488b83aa0983878101be772f9306a4b807727fff4daf943a24" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.540887 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2p7jp"] Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.546809 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2p7jp"] Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.547689 4880 scope.go:117] "RemoveContainer" containerID="e2a3aad412012ae55343cba5a3524a701d5c880169ae7c387336b7589c87a61c" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.566212 4880 scope.go:117] "RemoveContainer" containerID="c9411299b8c665d04606486e11555f55f58dc074ee7266ba0d58845b082e6c5f" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.586063 4880 scope.go:117] "RemoveContainer" containerID="27e4b33b5905857fdc8d2550542184f9e7ba78c8daa3994bed653587146a975f" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.604197 4880 scope.go:117] "RemoveContainer" containerID="0add42d8dbc19123b79a342223191b323060271d817ad75c93c6dec4895c6d2b" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.626560 4880 scope.go:117] "RemoveContainer" containerID="c9664e9a033f2c82472a0f2527f383aea2f06b296be8d876413eb76b10e9593f" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.651650 4880 scope.go:117] "RemoveContainer" containerID="5dc5fa58c032c4c57ebc7e43ad6e0c78712917493a0249313da5b35c25318203" Jan 26 16:05:38 crc kubenswrapper[4880]: I0126 16:05:38.670449 4880 scope.go:117] "RemoveContainer" containerID="9e849269649e61cd2d101f5b94692adee5480f51ef1889ea4c2f9daba19010ba" Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.480977 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"c4e2a6a42c87d1d912e803fc34f4ad3ec7a653378d8916a6201514e59f583ee1"} Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.481303 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"827272cccd7392b34cbabae30cacf8a218140494f890f2714975caf522db7ce8"} Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.481322 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"86e326e96d1fbc08b33a9b32ca9cd05b42077dab85d5e79fa32534d192ff37e2"} Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.481338 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"98126d013668c7c752efc98a7320233574da41757afe44e26c73e385f7ae76b4"} Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.481350 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"b0eb2f6909975779f87acfc5611b91aa64fff73fbacfaadd5780ffbe15f3d85c"} Jan 26 16:05:39 crc kubenswrapper[4880]: I0126 16:05:39.481361 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"e5bdb41d7fd295c8bbf6b5141a926085b555b7cc911eb4513c75dd5a749bf698"} Jan 26 16:05:40 crc kubenswrapper[4880]: I0126 16:05:40.220568 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c377ad66-a182-4cbf-9b29-253f3e547118" path="/var/lib/kubelet/pods/c377ad66-a182-4cbf-9b29-253f3e547118/volumes" Jan 26 16:05:43 crc kubenswrapper[4880]: I0126 16:05:43.510806 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"c48d06faeb7311a131263ec7c95d211f7b30c880194e981e39dc8b0fe32f6e47"} Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.527821 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" event={"ID":"d85f3c59-f9e9-4a12-8039-f9396436a803","Type":"ContainerStarted","Data":"14982b66a06d7ed00de2d7258728901c7f869a733b30fac240dffbc6b27f989f"} Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.528291 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.528312 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.560128 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.566114 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:45 crc kubenswrapper[4880]: I0126 16:05:45.567371 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" podStartSLOduration=8.567346527 podStartE2EDuration="8.567346527s" podCreationTimestamp="2026-01-26 16:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:05:45.565898751 +0000 UTC m=+671.051627458" watchObservedRunningTime="2026-01-26 16:05:45.567346527 +0000 UTC m=+671.053075234" Jan 26 16:05:46 crc kubenswrapper[4880]: I0126 16:05:46.534084 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:05:52 crc kubenswrapper[4880]: I0126 16:05:52.214178 4880 scope.go:117] "RemoveContainer" containerID="7963c3cc01b50b41810def1f17a183f3f03df1ec0d64b079bda8841b0e6f900c" Jan 26 16:05:52 crc kubenswrapper[4880]: I0126 16:05:52.626076 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-h6f8z_ddcd016a-fdad-4580-adaf-30451ac85f0e/kube-multus/1.log" Jan 26 16:05:52 crc kubenswrapper[4880]: I0126 16:05:52.626667 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-h6f8z" event={"ID":"ddcd016a-fdad-4580-adaf-30451ac85f0e","Type":"ContainerStarted","Data":"5f0c95eccb1b478bcb442565b208b60ef8f626af6f4ff2ab738f23e7a3f9bb29"} Jan 26 16:06:08 crc kubenswrapper[4880]: I0126 16:06:08.250315 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c6qlc" Jan 26 16:06:40 crc kubenswrapper[4880]: I0126 16:06:40.700502 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:06:40 crc kubenswrapper[4880]: I0126 16:06:40.701171 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.066164 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.067061 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7qhx" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="registry-server" containerID="cri-o://7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" gracePeriod=30 Jan 26 16:06:51 crc kubenswrapper[4880]: E0126 16:06:51.130797 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 is running failed: container process not found" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:06:51 crc kubenswrapper[4880]: E0126 16:06:51.131450 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 is running failed: container process not found" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:06:51 crc kubenswrapper[4880]: E0126 16:06:51.131663 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 is running failed: container process not found" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:06:51 crc kubenswrapper[4880]: E0126 16:06:51.131708 4880 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-t7qhx" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="registry-server" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.446709 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.645957 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities\") pod \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.646092 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content\") pod \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.646131 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtsrf\" (UniqueName: \"kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf\") pod \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\" (UID: \"6c79ba18-65a0-4ef8-87e9-1db08451a4e6\") " Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.647793 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities" (OuterVolumeSpecName: "utilities") pod "6c79ba18-65a0-4ef8-87e9-1db08451a4e6" (UID: "6c79ba18-65a0-4ef8-87e9-1db08451a4e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.652266 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf" (OuterVolumeSpecName: "kube-api-access-dtsrf") pod "6c79ba18-65a0-4ef8-87e9-1db08451a4e6" (UID: "6c79ba18-65a0-4ef8-87e9-1db08451a4e6"). InnerVolumeSpecName "kube-api-access-dtsrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.683274 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c79ba18-65a0-4ef8-87e9-1db08451a4e6" (UID: "6c79ba18-65a0-4ef8-87e9-1db08451a4e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.747969 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.748034 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:06:51 crc kubenswrapper[4880]: I0126 16:06:51.748054 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtsrf\" (UniqueName: \"kubernetes.io/projected/6c79ba18-65a0-4ef8-87e9-1db08451a4e6-kube-api-access-dtsrf\") on node \"crc\" DevicePath \"\"" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.034020 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7qhx" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.034074 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerDied","Data":"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805"} Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.034012 4880 generic.go:334] "Generic (PLEG): container finished" podID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" exitCode=0 Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.034155 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7qhx" event={"ID":"6c79ba18-65a0-4ef8-87e9-1db08451a4e6","Type":"ContainerDied","Data":"217e35f34cf348e472b5289f0377bbb10769a8ea301539d27204f184e0223a2c"} Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.034193 4880 scope.go:117] "RemoveContainer" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.055333 4880 scope.go:117] "RemoveContainer" containerID="e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.080966 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.085222 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7qhx"] Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.222711 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" path="/var/lib/kubelet/pods/6c79ba18-65a0-4ef8-87e9-1db08451a4e6/volumes" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.226482 4880 scope.go:117] "RemoveContainer" containerID="298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.252832 4880 scope.go:117] "RemoveContainer" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" Jan 26 16:06:52 crc kubenswrapper[4880]: E0126 16:06:52.253526 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805\": container with ID starting with 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 not found: ID does not exist" containerID="7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.253599 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805"} err="failed to get container status \"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805\": rpc error: code = NotFound desc = could not find container \"7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805\": container with ID starting with 7a61c60b53ddd7cb4dc621a6cb88f602b8b978c919b43c811df4cb1a19d01805 not found: ID does not exist" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.253652 4880 scope.go:117] "RemoveContainer" containerID="e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f" Jan 26 16:06:52 crc kubenswrapper[4880]: E0126 16:06:52.256604 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f\": container with ID starting with e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f not found: ID does not exist" containerID="e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.256650 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f"} err="failed to get container status \"e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f\": rpc error: code = NotFound desc = could not find container \"e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f\": container with ID starting with e5fb6653aab28bfc266ee1ca4e0482ca2d908b1499aa4be339431a425b68a44f not found: ID does not exist" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.256682 4880 scope.go:117] "RemoveContainer" containerID="298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b" Jan 26 16:06:52 crc kubenswrapper[4880]: E0126 16:06:52.257065 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b\": container with ID starting with 298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b not found: ID does not exist" containerID="298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b" Jan 26 16:06:52 crc kubenswrapper[4880]: I0126 16:06:52.257133 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b"} err="failed to get container status \"298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b\": rpc error: code = NotFound desc = could not find container \"298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b\": container with ID starting with 298133b52a650aebf747b7fb1a2c2ada1b2097de9461398b630240217e16c90b not found: ID does not exist" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.547225 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2"] Jan 26 16:06:55 crc kubenswrapper[4880]: E0126 16:06:55.547878 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="extract-utilities" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.547897 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="extract-utilities" Jan 26 16:06:55 crc kubenswrapper[4880]: E0126 16:06:55.547921 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="registry-server" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.547929 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="registry-server" Jan 26 16:06:55 crc kubenswrapper[4880]: E0126 16:06:55.547944 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="extract-content" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.547954 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="extract-content" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.548150 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c79ba18-65a0-4ef8-87e9-1db08451a4e6" containerName="registry-server" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.549289 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.552405 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.566454 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2"] Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.630021 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.630088 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzhw7\" (UniqueName: \"kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.630257 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.754114 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.754187 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzhw7\" (UniqueName: \"kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.754217 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.754914 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.755053 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.777067 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzhw7\" (UniqueName: \"kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:55 crc kubenswrapper[4880]: I0126 16:06:55.869961 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:06:56 crc kubenswrapper[4880]: I0126 16:06:56.139104 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2"] Jan 26 16:06:57 crc kubenswrapper[4880]: I0126 16:06:57.072529 4880 generic.go:334] "Generic (PLEG): container finished" podID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerID="0f49a5780862ba3e49746064a4a88238eb9022fd8b364828ba09e0f9f5eb9f4c" exitCode=0 Jan 26 16:06:57 crc kubenswrapper[4880]: I0126 16:06:57.072581 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerDied","Data":"0f49a5780862ba3e49746064a4a88238eb9022fd8b364828ba09e0f9f5eb9f4c"} Jan 26 16:06:57 crc kubenswrapper[4880]: I0126 16:06:57.072628 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerStarted","Data":"22312764b77d7618112117e16db8e9347afafb516145e8dcc2acad1c2c8678a7"} Jan 26 16:06:57 crc kubenswrapper[4880]: I0126 16:06:57.075186 4880 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 16:06:59 crc kubenswrapper[4880]: I0126 16:06:59.085192 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerStarted","Data":"256cd5d49caaeb876e498bf6adc3b947fe55640fd38da336c9242d13bc54825a"} Jan 26 16:07:00 crc kubenswrapper[4880]: I0126 16:07:00.094226 4880 generic.go:334] "Generic (PLEG): container finished" podID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerID="256cd5d49caaeb876e498bf6adc3b947fe55640fd38da336c9242d13bc54825a" exitCode=0 Jan 26 16:07:00 crc kubenswrapper[4880]: I0126 16:07:00.094272 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerDied","Data":"256cd5d49caaeb876e498bf6adc3b947fe55640fd38da336c9242d13bc54825a"} Jan 26 16:07:01 crc kubenswrapper[4880]: I0126 16:07:01.103110 4880 generic.go:334] "Generic (PLEG): container finished" podID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerID="d0f97046c1c56c867e0a4918d9154147e0c92bbd955d02538201c9a5c00ed9fb" exitCode=0 Jan 26 16:07:01 crc kubenswrapper[4880]: I0126 16:07:01.103153 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerDied","Data":"d0f97046c1c56c867e0a4918d9154147e0c92bbd955d02538201c9a5c00ed9fb"} Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.333494 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.343719 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle\") pod \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.343765 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzhw7\" (UniqueName: \"kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7\") pod \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.343830 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util\") pod \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\" (UID: \"7bcbfaec-4cc2-425e-be37-1556c5f92fdb\") " Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.346993 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle" (OuterVolumeSpecName: "bundle") pod "7bcbfaec-4cc2-425e-be37-1556c5f92fdb" (UID: "7bcbfaec-4cc2-425e-be37-1556c5f92fdb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.357856 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util" (OuterVolumeSpecName: "util") pod "7bcbfaec-4cc2-425e-be37-1556c5f92fdb" (UID: "7bcbfaec-4cc2-425e-be37-1556c5f92fdb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.358584 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7" (OuterVolumeSpecName: "kube-api-access-mzhw7") pod "7bcbfaec-4cc2-425e-be37-1556c5f92fdb" (UID: "7bcbfaec-4cc2-425e-be37-1556c5f92fdb"). InnerVolumeSpecName "kube-api-access-mzhw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.445775 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzhw7\" (UniqueName: \"kubernetes.io/projected/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-kube-api-access-mzhw7\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.445842 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:02 crc kubenswrapper[4880]: I0126 16:07:02.445862 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bcbfaec-4cc2-425e-be37-1556c5f92fdb-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:03 crc kubenswrapper[4880]: I0126 16:07:03.116313 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" event={"ID":"7bcbfaec-4cc2-425e-be37-1556c5f92fdb","Type":"ContainerDied","Data":"22312764b77d7618112117e16db8e9347afafb516145e8dcc2acad1c2c8678a7"} Jan 26 16:07:03 crc kubenswrapper[4880]: I0126 16:07:03.116717 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22312764b77d7618112117e16db8e9347afafb516145e8dcc2acad1c2c8678a7" Jan 26 16:07:03 crc kubenswrapper[4880]: I0126 16:07:03.116403 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.325183 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr"] Jan 26 16:07:04 crc kubenswrapper[4880]: E0126 16:07:04.325536 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="extract" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.325556 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="extract" Jan 26 16:07:04 crc kubenswrapper[4880]: E0126 16:07:04.325581 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="pull" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.325597 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="pull" Jan 26 16:07:04 crc kubenswrapper[4880]: E0126 16:07:04.325610 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="util" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.325619 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="util" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.325863 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bcbfaec-4cc2-425e-be37-1556c5f92fdb" containerName="extract" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.327006 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.331607 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.343201 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.368653 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.368720 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdc9m\" (UniqueName: \"kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.368973 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.470663 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.470764 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.470806 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdc9m\" (UniqueName: \"kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.471733 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.472485 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.491530 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdc9m\" (UniqueName: \"kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.650158 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.740060 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.747713 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.750320 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.777262 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.777336 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfxsq\" (UniqueName: \"kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.777383 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.881719 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.881856 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.881901 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfxsq\" (UniqueName: \"kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.882693 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.883025 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.930826 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.935661 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.950080 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.982286 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfxsq\" (UniqueName: \"kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.984809 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.986978 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.987034 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjghd\" (UniqueName: \"kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:04 crc kubenswrapper[4880]: I0126 16:07:04.987097 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.078905 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.091943 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjghd\" (UniqueName: \"kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.092016 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.092083 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.093125 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.093243 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.112658 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjghd\" (UniqueName: \"kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd\") pod \"redhat-operators-nr9fb\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.125864 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerStarted","Data":"f73e4233bd04ddf07e8d69bce98a5c71e8d23a432bce22f8d83ef2d797ec49fa"} Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.297319 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k"] Jan 26 16:07:05 crc kubenswrapper[4880]: W0126 16:07:05.300584 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d6d78ab_29ae_4468_9a9d_9e44ec38278c.slice/crio-3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa WatchSource:0}: Error finding container 3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa: Status 404 returned error can't find the container with id 3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.314532 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.347008 4880 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 16:07:05 crc kubenswrapper[4880]: I0126 16:07:05.575418 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:05 crc kubenswrapper[4880]: W0126 16:07:05.636501 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22383e0c_b3c0_4159_8edc_63e71175bff0.slice/crio-b2ee630d3ce38d9630ae7951b58df9442b413b67857cdb61d3c08b51467a536d WatchSource:0}: Error finding container b2ee630d3ce38d9630ae7951b58df9442b413b67857cdb61d3c08b51467a536d: Status 404 returned error can't find the container with id b2ee630d3ce38d9630ae7951b58df9442b413b67857cdb61d3c08b51467a536d Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.132448 4880 generic.go:334] "Generic (PLEG): container finished" podID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerID="b15570761a59075e133bcd3143296dde3fd193278d640cf13d2e16a1f2434951" exitCode=0 Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.132593 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerDied","Data":"b15570761a59075e133bcd3143296dde3fd193278d640cf13d2e16a1f2434951"} Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.132631 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerStarted","Data":"b2ee630d3ce38d9630ae7951b58df9442b413b67857cdb61d3c08b51467a536d"} Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.137258 4880 generic.go:334] "Generic (PLEG): container finished" podID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerID="3fa04279cc5eb21cfa92a03b46ba5d24718e2e7f3d8ef2d92c6b392c2d3eff20" exitCode=0 Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.137344 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerDied","Data":"3fa04279cc5eb21cfa92a03b46ba5d24718e2e7f3d8ef2d92c6b392c2d3eff20"} Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.137395 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerStarted","Data":"3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa"} Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.141124 4880 generic.go:334] "Generic (PLEG): container finished" podID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerID="b9408edd63e6bde9cd0107d44f29c9d921ba45c49458ac1bf403416e9e93dfe7" exitCode=0 Jan 26 16:07:06 crc kubenswrapper[4880]: I0126 16:07:06.141417 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerDied","Data":"b9408edd63e6bde9cd0107d44f29c9d921ba45c49458ac1bf403416e9e93dfe7"} Jan 26 16:07:08 crc kubenswrapper[4880]: I0126 16:07:08.156588 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerStarted","Data":"71ba7871322ff1fd1c315301c6b87643e75753c50e907b3c119f09b6469805c9"} Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.188780 4880 generic.go:334] "Generic (PLEG): container finished" podID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerID="71ba7871322ff1fd1c315301c6b87643e75753c50e907b3c119f09b6469805c9" exitCode=0 Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.189878 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerDied","Data":"71ba7871322ff1fd1c315301c6b87643e75753c50e907b3c119f09b6469805c9"} Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.310757 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.311842 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.341701 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.456343 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.456429 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blrnw\" (UniqueName: \"kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.456502 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.558138 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.558512 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.558650 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blrnw\" (UniqueName: \"kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.558687 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.558948 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.586752 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blrnw\" (UniqueName: \"kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw\") pod \"certified-operators-wdsvs\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:09 crc kubenswrapper[4880]: I0126 16:07:09.632062 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:10 crc kubenswrapper[4880]: I0126 16:07:10.700242 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:07:10 crc kubenswrapper[4880]: I0126 16:07:10.700318 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:07:12 crc kubenswrapper[4880]: I0126 16:07:12.464070 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.218699 4880 generic.go:334] "Generic (PLEG): container finished" podID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerID="7b26affc2a57e5f913720674801d756c75a56c62e3b3cef1415a66fc6f4f511b" exitCode=0 Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.218801 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerDied","Data":"7b26affc2a57e5f913720674801d756c75a56c62e3b3cef1415a66fc6f4f511b"} Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.221955 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerStarted","Data":"64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d"} Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.223800 4880 generic.go:334] "Generic (PLEG): container finished" podID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerID="51c03174209927342eaf0a6c5f2091fba7b819e21920a5d8982d60b505e2f3ae" exitCode=0 Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.223854 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerDied","Data":"51c03174209927342eaf0a6c5f2091fba7b819e21920a5d8982d60b505e2f3ae"} Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.223887 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerStarted","Data":"db4eb56f5e319d5d4258aa0e67f7f2a21c7f3df1ef213a3246726a68b0663346"} Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.225377 4880 generic.go:334] "Generic (PLEG): container finished" podID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerID="0d6b4877b3689255a9fe1495407d719dfc4d04e97a8b4067f3e87ddeb2264aad" exitCode=0 Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.225403 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerDied","Data":"0d6b4877b3689255a9fe1495407d719dfc4d04e97a8b4067f3e87ddeb2264aad"} Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.331615 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.337908 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.340629 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.340771 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.344089 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.344236 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-n5n9x" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.354272 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nr9fb" podStartSLOduration=3.026033009 podStartE2EDuration="9.354221163s" podCreationTimestamp="2026-01-26 16:07:04 +0000 UTC" firstStartedPulling="2026-01-26 16:07:06.134483547 +0000 UTC m=+751.620212254" lastFinishedPulling="2026-01-26 16:07:12.462671701 +0000 UTC m=+757.948400408" observedRunningTime="2026-01-26 16:07:13.347682169 +0000 UTC m=+758.833410876" watchObservedRunningTime="2026-01-26 16:07:13.354221163 +0000 UTC m=+758.839949870" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.422558 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.423394 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.426211 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-l6dcc" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.426708 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.434899 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.444991 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.445889 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.470810 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.513940 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf679\" (UniqueName: \"kubernetes.io/projected/2bf4440d-678a-4d07-95aa-e7aed961c47c-kube-api-access-wf679\") pod \"obo-prometheus-operator-68bc856cb9-l4qc7\" (UID: \"2bf4440d-678a-4d07-95aa-e7aed961c47c\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.615365 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.615464 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.615498 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.615561 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.615740 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf679\" (UniqueName: \"kubernetes.io/projected/2bf4440d-678a-4d07-95aa-e7aed961c47c-kube-api-access-wf679\") pod \"obo-prometheus-operator-68bc856cb9-l4qc7\" (UID: \"2bf4440d-678a-4d07-95aa-e7aed961c47c\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.654737 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf679\" (UniqueName: \"kubernetes.io/projected/2bf4440d-678a-4d07-95aa-e7aed961c47c-kube-api-access-wf679\") pod \"obo-prometheus-operator-68bc856cb9-l4qc7\" (UID: \"2bf4440d-678a-4d07-95aa-e7aed961c47c\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.674792 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.682176 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-t4mkq"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.683308 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:13 crc kubenswrapper[4880]: W0126 16:07:13.690214 4880 reflector.go:561] object-"openshift-operators"/"observability-operator-sa-dockercfg-nwmfq": failed to list *v1.Secret: secrets "observability-operator-sa-dockercfg-nwmfq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Jan 26 16:07:13 crc kubenswrapper[4880]: E0126 16:07:13.690314 4880 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"observability-operator-sa-dockercfg-nwmfq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"observability-operator-sa-dockercfg-nwmfq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 16:07:13 crc kubenswrapper[4880]: W0126 16:07:13.691016 4880 reflector.go:561] object-"openshift-operators"/"observability-operator-tls": failed to list *v1.Secret: secrets "observability-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Jan 26 16:07:13 crc kubenswrapper[4880]: E0126 16:07:13.691161 4880 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"observability-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"observability-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.715270 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-t4mkq"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.716942 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.716979 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.716997 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.717018 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.722186 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.722257 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.722725 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48492eaa-8a65-43ee-8658-917a89a5be96-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm\" (UID: \"48492eaa-8a65-43ee-8658-917a89a5be96\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.726407 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6d2ecd45-4798-48f7-a852-6007f00a1720-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-664d6f6879-b25tf\" (UID: \"6d2ecd45-4798-48f7-a852-6007f00a1720\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.741824 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.763477 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.819090 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3f672eb-d7d9-4316-a911-da9f76df698a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.819179 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbqh\" (UniqueName: \"kubernetes.io/projected/b3f672eb-d7d9-4316-a911-da9f76df698a-kube-api-access-gxbqh\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.830905 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dwwh5"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.831904 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.833987 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-c2rcc" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.840643 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dwwh5"] Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.919989 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbqh\" (UniqueName: \"kubernetes.io/projected/b3f672eb-d7d9-4316-a911-da9f76df698a-kube-api-access-gxbqh\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:13 crc kubenswrapper[4880]: I0126 16:07:13.920316 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3f672eb-d7d9-4316-a911-da9f76df698a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.115140 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5rr7\" (UniqueName: \"kubernetes.io/projected/611b80dc-f8d4-4043-ae3b-974b03133bb3-kube-api-access-x5rr7\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.115228 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/611b80dc-f8d4-4043-ae3b-974b03133bb3-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.128318 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbqh\" (UniqueName: \"kubernetes.io/projected/b3f672eb-d7d9-4316-a911-da9f76df698a-kube-api-access-gxbqh\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.270666 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5rr7\" (UniqueName: \"kubernetes.io/projected/611b80dc-f8d4-4043-ae3b-974b03133bb3-kube-api-access-x5rr7\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.270753 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/611b80dc-f8d4-4043-ae3b-974b03133bb3-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.272048 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/611b80dc-f8d4-4043-ae3b-974b03133bb3-openshift-service-ca\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.308517 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5rr7\" (UniqueName: \"kubernetes.io/projected/611b80dc-f8d4-4043-ae3b-974b03133bb3-kube-api-access-x5rr7\") pod \"perses-operator-5bf474d74f-dwwh5\" (UID: \"611b80dc-f8d4-4043-ae3b-974b03133bb3\") " pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.316292 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerStarted","Data":"6e772ecf73c910edd7326cc225b2209630995aefea118245b52dfb989ec9836d"} Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.331162 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerStarted","Data":"9511737dc9a287ce5b83bd4f42e786db5fa4bc086ae3e89bd64a151cfe0526eb"} Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.356990 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" podStartSLOduration=4.327369837 podStartE2EDuration="10.356968562s" podCreationTimestamp="2026-01-26 16:07:04 +0000 UTC" firstStartedPulling="2026-01-26 16:07:06.142723534 +0000 UTC m=+751.628452241" lastFinishedPulling="2026-01-26 16:07:12.172322259 +0000 UTC m=+757.658050966" observedRunningTime="2026-01-26 16:07:14.351342331 +0000 UTC m=+759.837071038" watchObservedRunningTime="2026-01-26 16:07:14.356968562 +0000 UTC m=+759.842697259" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.378620 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" podStartSLOduration=4.346095468 podStartE2EDuration="10.378597886s" podCreationTimestamp="2026-01-26 16:07:04 +0000 UTC" firstStartedPulling="2026-01-26 16:07:06.138789375 +0000 UTC m=+751.624518082" lastFinishedPulling="2026-01-26 16:07:12.171291793 +0000 UTC m=+757.657020500" observedRunningTime="2026-01-26 16:07:14.376857352 +0000 UTC m=+759.862586079" watchObservedRunningTime="2026-01-26 16:07:14.378597886 +0000 UTC m=+759.864326593" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.466840 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.489424 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm"] Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.556270 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.603648 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7"] Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.648995 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf"] Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.674577 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3f672eb-d7d9-4316-a911-da9f76df698a-observability-operator-tls\") pod \"observability-operator-59bdc8b94-t4mkq\" (UID: \"b3f672eb-d7d9-4316-a911-da9f76df698a\") " pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:14 crc kubenswrapper[4880]: I0126 16:07:14.846816 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-dwwh5"] Jan 26 16:07:14 crc kubenswrapper[4880]: W0126 16:07:14.963251 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod611b80dc_f8d4_4043_ae3b_974b03133bb3.slice/crio-f66af427037c8f8d6af92971f4e97e85f413089899bfebce6366bbba723a2bd1 WatchSource:0}: Error finding container f66af427037c8f8d6af92971f4e97e85f413089899bfebce6366bbba723a2bd1: Status 404 returned error can't find the container with id f66af427037c8f8d6af92971f4e97e85f413089899bfebce6366bbba723a2bd1 Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.115849 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-nwmfq" Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.123708 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.314767 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.314831 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.340269 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" event={"ID":"611b80dc-f8d4-4043-ae3b-974b03133bb3","Type":"ContainerStarted","Data":"f66af427037c8f8d6af92971f4e97e85f413089899bfebce6366bbba723a2bd1"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.343787 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-t4mkq"] Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.348800 4880 generic.go:334] "Generic (PLEG): container finished" podID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerID="6e772ecf73c910edd7326cc225b2209630995aefea118245b52dfb989ec9836d" exitCode=0 Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.348891 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerDied","Data":"6e772ecf73c910edd7326cc225b2209630995aefea118245b52dfb989ec9836d"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.351041 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" event={"ID":"2bf4440d-678a-4d07-95aa-e7aed961c47c","Type":"ContainerStarted","Data":"9e8f36cde48a9a318502cd4c0e4916163c1abb00717f7144af0bb3d2c670f734"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.352317 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" event={"ID":"6d2ecd45-4798-48f7-a852-6007f00a1720","Type":"ContainerStarted","Data":"d882e5580fbd6c650fd0d6240628762530c065720eb1836ad6613c23c612c983"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.353503 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" event={"ID":"48492eaa-8a65-43ee-8658-917a89a5be96","Type":"ContainerStarted","Data":"0cdd1c6c3ad757884dc4e2cf6238a74090288dd883a013e3413232a7e08217a3"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.356385 4880 generic.go:334] "Generic (PLEG): container finished" podID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerID="f561de891e1ba7afa87d8babad300a484d3201b5bb4cd881e47f5f84f233a2bd" exitCode=0 Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.356482 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerDied","Data":"f561de891e1ba7afa87d8babad300a484d3201b5bb4cd881e47f5f84f233a2bd"} Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.360855 4880 generic.go:334] "Generic (PLEG): container finished" podID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerID="9511737dc9a287ce5b83bd4f42e786db5fa4bc086ae3e89bd64a151cfe0526eb" exitCode=0 Jan 26 16:07:15 crc kubenswrapper[4880]: I0126 16:07:15.362312 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerDied","Data":"9511737dc9a287ce5b83bd4f42e786db5fa4bc086ae3e89bd64a151cfe0526eb"} Jan 26 16:07:15 crc kubenswrapper[4880]: W0126 16:07:15.375476 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3f672eb_d7d9_4316_a911_da9f76df698a.slice/crio-030785137bac70dd5edca3e02103154f75da16cfd2cf49125a4d5d13295f56e5 WatchSource:0}: Error finding container 030785137bac70dd5edca3e02103154f75da16cfd2cf49125a4d5d13295f56e5: Status 404 returned error can't find the container with id 030785137bac70dd5edca3e02103154f75da16cfd2cf49125a4d5d13295f56e5 Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.368991 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nr9fb" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" probeResult="failure" output=< Jan 26 16:07:16 crc kubenswrapper[4880]: timeout: failed to connect service ":50051" within 1s Jan 26 16:07:16 crc kubenswrapper[4880]: > Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.385428 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" event={"ID":"b3f672eb-d7d9-4316-a911-da9f76df698a","Type":"ContainerStarted","Data":"030785137bac70dd5edca3e02103154f75da16cfd2cf49125a4d5d13295f56e5"} Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.403501 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerStarted","Data":"b5bae48554e3a82c955e626618234e499567ce4e84f2c74f29f8343d9218e575"} Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.430169 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wdsvs" podStartSLOduration=4.856097206 podStartE2EDuration="7.430152823s" podCreationTimestamp="2026-01-26 16:07:09 +0000 UTC" firstStartedPulling="2026-01-26 16:07:13.238134314 +0000 UTC m=+758.723863021" lastFinishedPulling="2026-01-26 16:07:15.812189931 +0000 UTC m=+761.297918638" observedRunningTime="2026-01-26 16:07:16.428033789 +0000 UTC m=+761.913762496" watchObservedRunningTime="2026-01-26 16:07:16.430152823 +0000 UTC m=+761.915881550" Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.791592 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.897401 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.921900 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdc9m\" (UniqueName: \"kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m\") pod \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.921981 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle\") pod \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.922114 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util\") pod \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\" (UID: \"3d4820bd-1290-4cea-90fb-7ed5b43d5d55\") " Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.925005 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle" (OuterVolumeSpecName: "bundle") pod "3d4820bd-1290-4cea-90fb-7ed5b43d5d55" (UID: "3d4820bd-1290-4cea-90fb-7ed5b43d5d55"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.939360 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util" (OuterVolumeSpecName: "util") pod "3d4820bd-1290-4cea-90fb-7ed5b43d5d55" (UID: "3d4820bd-1290-4cea-90fb-7ed5b43d5d55"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:16 crc kubenswrapper[4880]: I0126 16:07:16.978612 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m" (OuterVolumeSpecName: "kube-api-access-cdc9m") pod "3d4820bd-1290-4cea-90fb-7ed5b43d5d55" (UID: "3d4820bd-1290-4cea-90fb-7ed5b43d5d55"). InnerVolumeSpecName "kube-api-access-cdc9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.025957 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util\") pod \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.026119 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfxsq\" (UniqueName: \"kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq\") pod \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.026215 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle\") pod \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\" (UID: \"4d6d78ab-29ae-4468-9a9d-9e44ec38278c\") " Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.026511 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.026535 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdc9m\" (UniqueName: \"kubernetes.io/projected/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-kube-api-access-cdc9m\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.026550 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d4820bd-1290-4cea-90fb-7ed5b43d5d55-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.027741 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle" (OuterVolumeSpecName: "bundle") pod "4d6d78ab-29ae-4468-9a9d-9e44ec38278c" (UID: "4d6d78ab-29ae-4468-9a9d-9e44ec38278c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.031462 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq" (OuterVolumeSpecName: "kube-api-access-jfxsq") pod "4d6d78ab-29ae-4468-9a9d-9e44ec38278c" (UID: "4d6d78ab-29ae-4468-9a9d-9e44ec38278c"). InnerVolumeSpecName "kube-api-access-jfxsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.046034 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util" (OuterVolumeSpecName: "util") pod "4d6d78ab-29ae-4468-9a9d-9e44ec38278c" (UID: "4d6d78ab-29ae-4468-9a9d-9e44ec38278c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.127453 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.127494 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.127504 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfxsq\" (UniqueName: \"kubernetes.io/projected/4d6d78ab-29ae-4468-9a9d-9e44ec38278c-kube-api-access-jfxsq\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.439615 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" event={"ID":"4d6d78ab-29ae-4468-9a9d-9e44ec38278c","Type":"ContainerDied","Data":"3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa"} Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.440251 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b67d7313e614bd63fcd7cbc8d80dc2114528c9818ddbbe184f33f61b8ba4aaa" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.440335 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.445663 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.451612 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr" event={"ID":"3d4820bd-1290-4cea-90fb-7ed5b43d5d55","Type":"ContainerDied","Data":"f73e4233bd04ddf07e8d69bce98a5c71e8d23a432bce22f8d83ef2d797ec49fa"} Jan 26 16:07:17 crc kubenswrapper[4880]: I0126 16:07:17.451676 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f73e4233bd04ddf07e8d69bce98a5c71e8d23a432bce22f8d83ef2d797ec49fa" Jan 26 16:07:19 crc kubenswrapper[4880]: I0126 16:07:19.632839 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:19 crc kubenswrapper[4880]: I0126 16:07:19.632937 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:19 crc kubenswrapper[4880]: I0126 16:07:19.702321 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:20 crc kubenswrapper[4880]: I0126 16:07:20.623242 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.774563 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-6c675c8ff8-rw5mq"] Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775481 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="util" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775500 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="util" Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775523 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="pull" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775533 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="pull" Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775544 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775553 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775571 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="util" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775581 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="util" Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775596 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775604 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: E0126 16:07:21.775615 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="pull" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775623 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="pull" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775762 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4820bd-1290-4cea-90fb-7ed5b43d5d55" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.775782 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6d78ab-29ae-4468-9a9d-9e44ec38278c" containerName="extract" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.776364 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.780103 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-fg2dl" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.780294 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.780547 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.780953 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.799727 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-6c675c8ff8-rw5mq"] Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.889245 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-webhook-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.889323 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-apiservice-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:21 crc kubenswrapper[4880]: I0126 16:07:21.889399 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6pq6\" (UniqueName: \"kubernetes.io/projected/d71a4bc1-9f17-437b-a25d-9edefb0ad989-kube-api-access-z6pq6\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.006820 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6pq6\" (UniqueName: \"kubernetes.io/projected/d71a4bc1-9f17-437b-a25d-9edefb0ad989-kube-api-access-z6pq6\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.007005 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-webhook-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.007049 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-apiservice-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.015683 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-webhook-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.019125 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d71a4bc1-9f17-437b-a25d-9edefb0ad989-apiservice-cert\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.025814 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6pq6\" (UniqueName: \"kubernetes.io/projected/d71a4bc1-9f17-437b-a25d-9edefb0ad989-kube-api-access-z6pq6\") pod \"elastic-operator-6c675c8ff8-rw5mq\" (UID: \"d71a4bc1-9f17-437b-a25d-9edefb0ad989\") " pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.080350 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.124829 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" Jan 26 16:07:22 crc kubenswrapper[4880]: I0126 16:07:22.509740 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wdsvs" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="registry-server" containerID="cri-o://b5bae48554e3a82c955e626618234e499567ce4e84f2c74f29f8343d9218e575" gracePeriod=2 Jan 26 16:07:23 crc kubenswrapper[4880]: I0126 16:07:23.528392 4880 generic.go:334] "Generic (PLEG): container finished" podID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerID="b5bae48554e3a82c955e626618234e499567ce4e84f2c74f29f8343d9218e575" exitCode=0 Jan 26 16:07:23 crc kubenswrapper[4880]: I0126 16:07:23.528768 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerDied","Data":"b5bae48554e3a82c955e626618234e499567ce4e84f2c74f29f8343d9218e575"} Jan 26 16:07:25 crc kubenswrapper[4880]: I0126 16:07:25.508751 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:25 crc kubenswrapper[4880]: I0126 16:07:25.569524 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.227906 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.388977 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content\") pod \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.389072 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blrnw\" (UniqueName: \"kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw\") pod \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.389117 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities\") pod \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\" (UID: \"dc3dbccd-8579-449a-bbf3-9b09b9afeab9\") " Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.390418 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities" (OuterVolumeSpecName: "utilities") pod "dc3dbccd-8579-449a-bbf3-9b09b9afeab9" (UID: "dc3dbccd-8579-449a-bbf3-9b09b9afeab9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.411716 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw" (OuterVolumeSpecName: "kube-api-access-blrnw") pod "dc3dbccd-8579-449a-bbf3-9b09b9afeab9" (UID: "dc3dbccd-8579-449a-bbf3-9b09b9afeab9"). InnerVolumeSpecName "kube-api-access-blrnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.440145 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc3dbccd-8579-449a-bbf3-9b09b9afeab9" (UID: "dc3dbccd-8579-449a-bbf3-9b09b9afeab9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.490554 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.490609 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.490625 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blrnw\" (UniqueName: \"kubernetes.io/projected/dc3dbccd-8579-449a-bbf3-9b09b9afeab9-kube-api-access-blrnw\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.831244 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdsvs" event={"ID":"dc3dbccd-8579-449a-bbf3-9b09b9afeab9","Type":"ContainerDied","Data":"db4eb56f5e319d5d4258aa0e67f7f2a21c7f3df1ef213a3246726a68b0663346"} Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.831351 4880 scope.go:117] "RemoveContainer" containerID="b5bae48554e3a82c955e626618234e499567ce4e84f2c74f29f8343d9218e575" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.831577 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdsvs" Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.875420 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:29 crc kubenswrapper[4880]: I0126 16:07:29.882617 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wdsvs"] Jan 26 16:07:30 crc kubenswrapper[4880]: I0126 16:07:30.105401 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:30 crc kubenswrapper[4880]: I0126 16:07:30.105789 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nr9fb" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" containerID="cri-o://64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" gracePeriod=2 Jan 26 16:07:30 crc kubenswrapper[4880]: I0126 16:07:30.613245 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" path="/var/lib/kubelet/pods/dc3dbccd-8579-449a-bbf3-9b09b9afeab9/volumes" Jan 26 16:07:31 crc kubenswrapper[4880]: I0126 16:07:31.942676 4880 generic.go:334] "Generic (PLEG): container finished" podID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerID="64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" exitCode=0 Jan 26 16:07:31 crc kubenswrapper[4880]: I0126 16:07:31.942751 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerDied","Data":"64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d"} Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.064337 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns"] Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.065002 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="extract-content" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.065025 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="extract-content" Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.065047 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="registry-server" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.065056 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="registry-server" Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.065075 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="extract-utilities" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.065085 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="extract-utilities" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.065359 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3dbccd-8579-449a-bbf3-9b09b9afeab9" containerName="registry-server" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.066069 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.068657 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.068814 4880 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-hzpl8" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.071591 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.090515 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns"] Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.155782 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0c048a40-ed87-417f-8ec9-207753c038b6-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.155862 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpq8c\" (UniqueName: \"kubernetes.io/projected/0c048a40-ed87-417f-8ec9-207753c038b6-kube-api-access-vpq8c\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.257190 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpq8c\" (UniqueName: \"kubernetes.io/projected/0c048a40-ed87-417f-8ec9-207753c038b6-kube-api-access-vpq8c\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.257585 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0c048a40-ed87-417f-8ec9-207753c038b6-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.258056 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0c048a40-ed87-417f-8ec9-207753c038b6-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.281141 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpq8c\" (UniqueName: \"kubernetes.io/projected/0c048a40-ed87-417f-8ec9-207753c038b6-kube-api-access-vpq8c\") pod \"cert-manager-operator-controller-manager-5446d6888b-f2gns\" (UID: \"0c048a40-ed87-417f-8ec9-207753c038b6\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.316622 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d is running failed: container process not found" containerID="64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.316916 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d is running failed: container process not found" containerID="64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.317214 4880 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d is running failed: container process not found" containerID="64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 16:07:35 crc kubenswrapper[4880]: E0126 16:07:35.317254 4880 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-nr9fb" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" Jan 26 16:07:35 crc kubenswrapper[4880]: I0126 16:07:35.388870 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" Jan 26 16:07:36 crc kubenswrapper[4880]: E0126 16:07:36.357594 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8" Jan 26 16:07:36 crc kubenswrapper[4880]: E0126 16:07:36.357965 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5rr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5bf474d74f-dwwh5_openshift-operators(611b80dc-f8d4-4043-ae3b-974b03133bb3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 16:07:36 crc kubenswrapper[4880]: E0126 16:07:36.359220 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" podUID="611b80dc-f8d4-4043-ae3b-974b03133bb3" Jan 26 16:07:36 crc kubenswrapper[4880]: I0126 16:07:36.831043 4880 scope.go:117] "RemoveContainer" containerID="f561de891e1ba7afa87d8babad300a484d3201b5bb4cd881e47f5f84f233a2bd" Jan 26 16:07:36 crc kubenswrapper[4880]: I0126 16:07:36.928629 4880 scope.go:117] "RemoveContainer" containerID="51c03174209927342eaf0a6c5f2091fba7b819e21920a5d8982d60b505e2f3ae" Jan 26 16:07:37 crc kubenswrapper[4880]: E0126 16:07:37.023791 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8\\\"\"" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" podUID="611b80dc-f8d4-4043-ae3b-974b03133bb3" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.166907 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-6c675c8ff8-rw5mq"] Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.219517 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:37 crc kubenswrapper[4880]: W0126 16:07:37.236354 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd71a4bc1_9f17_437b_a25d_9edefb0ad989.slice/crio-f407dbdb51b9569d619a2ae1fc785f072e409c3db8ee5a62648d5614bcffd6e3 WatchSource:0}: Error finding container f407dbdb51b9569d619a2ae1fc785f072e409c3db8ee5a62648d5614bcffd6e3: Status 404 returned error can't find the container with id f407dbdb51b9569d619a2ae1fc785f072e409c3db8ee5a62648d5614bcffd6e3 Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.250625 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns"] Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.389697 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjghd\" (UniqueName: \"kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd\") pod \"22383e0c-b3c0-4159-8edc-63e71175bff0\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.389819 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content\") pod \"22383e0c-b3c0-4159-8edc-63e71175bff0\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.389882 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities\") pod \"22383e0c-b3c0-4159-8edc-63e71175bff0\" (UID: \"22383e0c-b3c0-4159-8edc-63e71175bff0\") " Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.392013 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities" (OuterVolumeSpecName: "utilities") pod "22383e0c-b3c0-4159-8edc-63e71175bff0" (UID: "22383e0c-b3c0-4159-8edc-63e71175bff0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.398650 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd" (OuterVolumeSpecName: "kube-api-access-hjghd") pod "22383e0c-b3c0-4159-8edc-63e71175bff0" (UID: "22383e0c-b3c0-4159-8edc-63e71175bff0"). InnerVolumeSpecName "kube-api-access-hjghd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.491657 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.491695 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjghd\" (UniqueName: \"kubernetes.io/projected/22383e0c-b3c0-4159-8edc-63e71175bff0-kube-api-access-hjghd\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.524148 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22383e0c-b3c0-4159-8edc-63e71175bff0" (UID: "22383e0c-b3c0-4159-8edc-63e71175bff0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:07:37 crc kubenswrapper[4880]: I0126 16:07:37.593629 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22383e0c-b3c0-4159-8edc-63e71175bff0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.030054 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" event={"ID":"b3f672eb-d7d9-4316-a911-da9f76df698a","Type":"ContainerStarted","Data":"676b46800de70ca57b0aabe341f09e7fafe1ce6aa1b61058358a9eb131037832"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.030377 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.032914 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" event={"ID":"d71a4bc1-9f17-437b-a25d-9edefb0ad989","Type":"ContainerStarted","Data":"f407dbdb51b9569d619a2ae1fc785f072e409c3db8ee5a62648d5614bcffd6e3"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.036656 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fb" event={"ID":"22383e0c-b3c0-4159-8edc-63e71175bff0","Type":"ContainerDied","Data":"b2ee630d3ce38d9630ae7951b58df9442b413b67857cdb61d3c08b51467a536d"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.036691 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fb" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.036733 4880 scope.go:117] "RemoveContainer" containerID="64e632850d36ea4cf7662e31b588d1f9125bb99f7412a735c5823d869a5d8a8d" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.038416 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" event={"ID":"0c048a40-ed87-417f-8ec9-207753c038b6","Type":"ContainerStarted","Data":"bd06190ea36efed626257f60f1ba0fa7b09c83105c1ca603cdc85f3e42425741"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.040588 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" event={"ID":"2bf4440d-678a-4d07-95aa-e7aed961c47c","Type":"ContainerStarted","Data":"b926acf96206d44f0965c107d621cd1045b2e86d9ccf696bba1a92ce44e8462f"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.042803 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" event={"ID":"6d2ecd45-4798-48f7-a852-6007f00a1720","Type":"ContainerStarted","Data":"ae25a7c09798b5b01e3451db191526fa15c7d32f132033aa6d6ecede5e270c60"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.044597 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" event={"ID":"48492eaa-8a65-43ee-8658-917a89a5be96","Type":"ContainerStarted","Data":"70ad68f479af1f6eea2e07af9c525226b7360a851cff732a0310879036deb95c"} Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.057744 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" podStartSLOduration=3.6014239569999997 podStartE2EDuration="25.057662281s" podCreationTimestamp="2026-01-26 16:07:13 +0000 UTC" firstStartedPulling="2026-01-26 16:07:15.383463978 +0000 UTC m=+760.869192685" lastFinishedPulling="2026-01-26 16:07:36.839702302 +0000 UTC m=+782.325431009" observedRunningTime="2026-01-26 16:07:38.050288205 +0000 UTC m=+783.536016912" watchObservedRunningTime="2026-01-26 16:07:38.057662281 +0000 UTC m=+783.543390998" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.091655 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-t4mkq" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.094021 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-b25tf" podStartSLOduration=2.97394214 podStartE2EDuration="25.094002658s" podCreationTimestamp="2026-01-26 16:07:13 +0000 UTC" firstStartedPulling="2026-01-26 16:07:14.699953828 +0000 UTC m=+760.185682535" lastFinishedPulling="2026-01-26 16:07:36.820014346 +0000 UTC m=+782.305743053" observedRunningTime="2026-01-26 16:07:38.091960357 +0000 UTC m=+783.577689064" watchObservedRunningTime="2026-01-26 16:07:38.094002658 +0000 UTC m=+783.579731355" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.125391 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm" podStartSLOduration=2.8053299320000002 podStartE2EDuration="25.125374841s" podCreationTimestamp="2026-01-26 16:07:13 +0000 UTC" firstStartedPulling="2026-01-26 16:07:14.519255524 +0000 UTC m=+760.004984231" lastFinishedPulling="2026-01-26 16:07:36.839300433 +0000 UTC m=+782.325029140" observedRunningTime="2026-01-26 16:07:38.122292712 +0000 UTC m=+783.608021429" watchObservedRunningTime="2026-01-26 16:07:38.125374841 +0000 UTC m=+783.611103538" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.150082 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-l4qc7" podStartSLOduration=3.001416255 podStartE2EDuration="25.150058763s" podCreationTimestamp="2026-01-26 16:07:13 +0000 UTC" firstStartedPulling="2026-01-26 16:07:14.690937931 +0000 UTC m=+760.176666638" lastFinishedPulling="2026-01-26 16:07:36.839580439 +0000 UTC m=+782.325309146" observedRunningTime="2026-01-26 16:07:38.148421072 +0000 UTC m=+783.634149779" watchObservedRunningTime="2026-01-26 16:07:38.150058763 +0000 UTC m=+783.635787470" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.203491 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.209989 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fb"] Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.227551 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" path="/var/lib/kubelet/pods/22383e0c-b3c0-4159-8edc-63e71175bff0/volumes" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.599973 4880 scope.go:117] "RemoveContainer" containerID="71ba7871322ff1fd1c315301c6b87643e75753c50e907b3c119f09b6469805c9" Jan 26 16:07:38 crc kubenswrapper[4880]: I0126 16:07:38.629390 4880 scope.go:117] "RemoveContainer" containerID="b15570761a59075e133bcd3143296dde3fd193278d640cf13d2e16a1f2434951" Jan 26 16:07:40 crc kubenswrapper[4880]: I0126 16:07:40.699733 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:07:40 crc kubenswrapper[4880]: I0126 16:07:40.700343 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:07:40 crc kubenswrapper[4880]: I0126 16:07:40.700400 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:07:40 crc kubenswrapper[4880]: I0126 16:07:40.701157 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:07:40 crc kubenswrapper[4880]: I0126 16:07:40.701224 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da" gracePeriod=600 Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.072664 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" event={"ID":"d71a4bc1-9f17-437b-a25d-9edefb0ad989","Type":"ContainerStarted","Data":"20b1d49255a2f20b634bfd92c90b81cc933612decbf9565b237f9425468d57be"} Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.077596 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da" exitCode=0 Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.077642 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da"} Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.077670 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56"} Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.077692 4880 scope.go:117] "RemoveContainer" containerID="2934124bf71d350f016db71086a4a29c87086032aa008fddfa1de4b976c7f585" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.094378 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-6c675c8ff8-rw5mq" podStartSLOduration=17.112790052 podStartE2EDuration="20.094336381s" podCreationTimestamp="2026-01-26 16:07:21 +0000 UTC" firstStartedPulling="2026-01-26 16:07:37.243186225 +0000 UTC m=+782.728914932" lastFinishedPulling="2026-01-26 16:07:40.224732554 +0000 UTC m=+785.710461261" observedRunningTime="2026-01-26 16:07:41.09430522 +0000 UTC m=+786.580033927" watchObservedRunningTime="2026-01-26 16:07:41.094336381 +0000 UTC m=+786.580065098" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.483498 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 16:07:41 crc kubenswrapper[4880]: E0126 16:07:41.484057 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="extract-content" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.484085 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="extract-content" Jan 26 16:07:41 crc kubenswrapper[4880]: E0126 16:07:41.484101 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.484109 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" Jan 26 16:07:41 crc kubenswrapper[4880]: E0126 16:07:41.484124 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="extract-utilities" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.484132 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="extract-utilities" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.484280 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="22383e0c-b3c0-4159-8edc-63e71175bff0" containerName="registry-server" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.485262 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.488748 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.489158 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.489535 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.489589 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.489642 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.489775 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-g5dxz" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.491237 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.492876 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.498520 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.501885 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.653965 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654025 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654056 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/fff0a88a-97cf-45a6-8de4-4568e40bac22-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654197 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654286 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654327 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654377 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654410 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654535 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654568 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654619 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654651 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654771 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654806 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.654858 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763403 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763483 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763521 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763600 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763637 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763661 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/fff0a88a-97cf-45a6-8de4-4568e40bac22-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763681 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763714 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763738 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763764 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763784 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763816 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763837 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763861 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.763884 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.764322 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.764409 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.764694 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.764910 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.765207 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.765394 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.765538 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.766522 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.768994 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.769102 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/fff0a88a-97cf-45a6-8de4-4568e40bac22-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.771706 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.779130 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.782461 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.783170 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.784596 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/fff0a88a-97cf-45a6-8de4-4568e40bac22-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"fff0a88a-97cf-45a6-8de4-4568e40bac22\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:41 crc kubenswrapper[4880]: I0126 16:07:41.807120 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:07:46 crc kubenswrapper[4880]: I0126 16:07:46.121500 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" event={"ID":"0c048a40-ed87-417f-8ec9-207753c038b6","Type":"ContainerStarted","Data":"04d8aad2b0612f64605cd7f935524014df243048eb144358470017b7625adf96"} Jan 26 16:07:46 crc kubenswrapper[4880]: I0126 16:07:46.138969 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-f2gns" podStartSLOduration=2.467002293 podStartE2EDuration="11.138942265s" podCreationTimestamp="2026-01-26 16:07:35 +0000 UTC" firstStartedPulling="2026-01-26 16:07:37.262603025 +0000 UTC m=+782.748331732" lastFinishedPulling="2026-01-26 16:07:45.934542997 +0000 UTC m=+791.420271704" observedRunningTime="2026-01-26 16:07:46.133264242 +0000 UTC m=+791.618992959" watchObservedRunningTime="2026-01-26 16:07:46.138942265 +0000 UTC m=+791.624670982" Jan 26 16:07:46 crc kubenswrapper[4880]: I0126 16:07:46.285540 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 16:07:46 crc kubenswrapper[4880]: W0126 16:07:46.288833 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfff0a88a_97cf_45a6_8de4_4568e40bac22.slice/crio-a3f23713a7aca9b134d68398b13919fb4ce98d73680859a6a1007ff76a0c635e WatchSource:0}: Error finding container a3f23713a7aca9b134d68398b13919fb4ce98d73680859a6a1007ff76a0c635e: Status 404 returned error can't find the container with id a3f23713a7aca9b134d68398b13919fb4ce98d73680859a6a1007ff76a0c635e Jan 26 16:07:47 crc kubenswrapper[4880]: I0126 16:07:47.129987 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"fff0a88a-97cf-45a6-8de4-4568e40bac22","Type":"ContainerStarted","Data":"a3f23713a7aca9b134d68398b13919fb4ce98d73680859a6a1007ff76a0c635e"} Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.407363 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-v28tf"] Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.409803 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.413580 4880 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-pd245" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.413893 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.414045 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.417855 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-v28tf"] Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.549648 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.550378 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5w6b\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-kube-api-access-z5w6b\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.651383 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5w6b\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-kube-api-access-z5w6b\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.651478 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.673201 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5w6b\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-kube-api-access-z5w6b\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.687328 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b957e4f4-7d0c-428c-bf37-4f4362d0e4bd-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-v28tf\" (UID: \"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:54 crc kubenswrapper[4880]: I0126 16:07:54.736562 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.807827 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-snnps"] Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.808902 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.810954 4880 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vdlmp" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.820741 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-snnps"] Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.874380 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.874494 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m6xl\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-kube-api-access-7m6xl\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.975912 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:55 crc kubenswrapper[4880]: I0126 16:07:55.976030 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m6xl\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-kube-api-access-7m6xl\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:56 crc kubenswrapper[4880]: I0126 16:07:56.010482 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m6xl\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-kube-api-access-7m6xl\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:56 crc kubenswrapper[4880]: I0126 16:07:56.059914 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-snnps\" (UID: \"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:07:56 crc kubenswrapper[4880]: I0126 16:07:56.130567 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.230285 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-snnps"] Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.234881 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-v28tf"] Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.454101 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"fff0a88a-97cf-45a6-8de4-4568e40bac22","Type":"ContainerStarted","Data":"9b31a6a11a76cd10b4bc0913c144ca19bd3d2ec0dff8804f5710d8fe8075ca67"} Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.456371 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" event={"ID":"611b80dc-f8d4-4043-ae3b-974b03133bb3","Type":"ContainerStarted","Data":"b991dc706b974003b3d04434a2c834fb34d7b794d7f2178ebb51541e844286ba"} Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.456670 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.458474 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" event={"ID":"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee","Type":"ContainerStarted","Data":"44ad5ee4e52b65b2628d0e627593c8bf50baa3ed7974896a9a4a166af4a3b14b"} Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.459963 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" event={"ID":"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd","Type":"ContainerStarted","Data":"04aac59db0924020c92be1e06f288ed38e4c9c3cb29f3a6134c7d575899267f9"} Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.520931 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" podStartSLOduration=2.773241942 podStartE2EDuration="48.520893883s" podCreationTimestamp="2026-01-26 16:07:13 +0000 UTC" firstStartedPulling="2026-01-26 16:07:14.966620485 +0000 UTC m=+760.452349192" lastFinishedPulling="2026-01-26 16:08:00.714272426 +0000 UTC m=+806.200001133" observedRunningTime="2026-01-26 16:08:01.515836915 +0000 UTC m=+807.001565622" watchObservedRunningTime="2026-01-26 16:08:01.520893883 +0000 UTC m=+807.006622590" Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.623779 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 16:08:01 crc kubenswrapper[4880]: I0126 16:08:01.668044 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 16:08:03 crc kubenswrapper[4880]: I0126 16:08:03.493999 4880 generic.go:334] "Generic (PLEG): container finished" podID="fff0a88a-97cf-45a6-8de4-4568e40bac22" containerID="9b31a6a11a76cd10b4bc0913c144ca19bd3d2ec0dff8804f5710d8fe8075ca67" exitCode=0 Jan 26 16:08:03 crc kubenswrapper[4880]: I0126 16:08:03.494090 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"fff0a88a-97cf-45a6-8de4-4568e40bac22","Type":"ContainerDied","Data":"9b31a6a11a76cd10b4bc0913c144ca19bd3d2ec0dff8804f5710d8fe8075ca67"} Jan 26 16:08:04 crc kubenswrapper[4880]: I0126 16:08:04.639758 4880 generic.go:334] "Generic (PLEG): container finished" podID="fff0a88a-97cf-45a6-8de4-4568e40bac22" containerID="8afb36c318c3d7b24900ee2ca84c71424444d6371c69fd53bc1f49cd789b4aeb" exitCode=0 Jan 26 16:08:04 crc kubenswrapper[4880]: I0126 16:08:04.639820 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"fff0a88a-97cf-45a6-8de4-4568e40bac22","Type":"ContainerDied","Data":"8afb36c318c3d7b24900ee2ca84c71424444d6371c69fd53bc1f49cd789b4aeb"} Jan 26 16:08:05 crc kubenswrapper[4880]: I0126 16:08:05.654133 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"fff0a88a-97cf-45a6-8de4-4568e40bac22","Type":"ContainerStarted","Data":"94de7d7b6bdd6c562cf0ed5f670e9ef92ddf5fb8713645778e19727c9b320eeb"} Jan 26 16:08:05 crc kubenswrapper[4880]: I0126 16:08:05.654390 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:08:05 crc kubenswrapper[4880]: I0126 16:08:05.702458 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=10.092400429 podStartE2EDuration="24.702417766s" podCreationTimestamp="2026-01-26 16:07:41 +0000 UTC" firstStartedPulling="2026-01-26 16:07:46.290669034 +0000 UTC m=+791.776397741" lastFinishedPulling="2026-01-26 16:08:00.900686371 +0000 UTC m=+806.386415078" observedRunningTime="2026-01-26 16:08:05.689236624 +0000 UTC m=+811.174965331" watchObservedRunningTime="2026-01-26 16:08:05.702417766 +0000 UTC m=+811.188146473" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.445695 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-fq6ff"] Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.447017 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.449462 4880 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-hljvk" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.462859 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-fq6ff"] Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.497702 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r94r7\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-kube-api-access-r94r7\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.497833 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-bound-sa-token\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.598820 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r94r7\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-kube-api-access-r94r7\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.598875 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-bound-sa-token\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.620053 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r94r7\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-kube-api-access-r94r7\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.620067 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e5326281-b231-4a2d-a1de-608c5815e977-bound-sa-token\") pod \"cert-manager-86cb77c54b-fq6ff\" (UID: \"e5326281-b231-4a2d-a1de-608c5815e977\") " pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.772814 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-fq6ff" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.842474 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-sz2ch"] Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.843584 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.852749 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-rtb7r" Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.862547 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-sz2ch"] Jan 26 16:08:08 crc kubenswrapper[4880]: I0126 16:08:08.903139 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jt5l\" (UniqueName: \"kubernetes.io/projected/f77203c1-a597-4711-aac7-0f20f88aef3f-kube-api-access-4jt5l\") pod \"infrawatch-operators-sz2ch\" (UID: \"f77203c1-a597-4711-aac7-0f20f88aef3f\") " pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:09 crc kubenswrapper[4880]: I0126 16:08:09.005512 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jt5l\" (UniqueName: \"kubernetes.io/projected/f77203c1-a597-4711-aac7-0f20f88aef3f-kube-api-access-4jt5l\") pod \"infrawatch-operators-sz2ch\" (UID: \"f77203c1-a597-4711-aac7-0f20f88aef3f\") " pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:09 crc kubenswrapper[4880]: I0126 16:08:09.030360 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jt5l\" (UniqueName: \"kubernetes.io/projected/f77203c1-a597-4711-aac7-0f20f88aef3f-kube-api-access-4jt5l\") pod \"infrawatch-operators-sz2ch\" (UID: \"f77203c1-a597-4711-aac7-0f20f88aef3f\") " pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:09 crc kubenswrapper[4880]: I0126 16:08:09.165615 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.059522 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-fq6ff"] Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.153920 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-sz2ch"] Jan 26 16:08:12 crc kubenswrapper[4880]: W0126 16:08:12.163792 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf77203c1_a597_4711_aac7_0f20f88aef3f.slice/crio-84d412b98ffd8a9eef92b536bb5b439dcfd3f2722f935bc688210eb5b9753be6 WatchSource:0}: Error finding container 84d412b98ffd8a9eef92b536bb5b439dcfd3f2722f935bc688210eb5b9753be6: Status 404 returned error can't find the container with id 84d412b98ffd8a9eef92b536bb5b439dcfd3f2722f935bc688210eb5b9753be6 Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.701103 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" event={"ID":"94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee","Type":"ContainerStarted","Data":"1d681c911a8f79e3c7652562e5c19b16689b21a0a4c84e95ae5a94c7562ab6d7"} Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.703321 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" event={"ID":"b957e4f4-7d0c-428c-bf37-4f4362d0e4bd","Type":"ContainerStarted","Data":"8ade02de5996f11f056933b24430dd79b253363ca83ee761f861db1b9d9d91a9"} Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.703446 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.704238 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-sz2ch" event={"ID":"f77203c1-a597-4711-aac7-0f20f88aef3f","Type":"ContainerStarted","Data":"84d412b98ffd8a9eef92b536bb5b439dcfd3f2722f935bc688210eb5b9753be6"} Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.705516 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-fq6ff" event={"ID":"e5326281-b231-4a2d-a1de-608c5815e977","Type":"ContainerStarted","Data":"8ef5c53abf7089631db7cb9f9ad871aeabdb0f739edc1bcc07fb22d7416f67a4"} Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.705581 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-fq6ff" event={"ID":"e5326281-b231-4a2d-a1de-608c5815e977","Type":"ContainerStarted","Data":"6f5566c0f17d10c90ca71697edb7d95b0c859167f0d74ca76fcfea83466b4db6"} Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.722408 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-snnps" podStartSLOduration=7.20345727 podStartE2EDuration="17.722387315s" podCreationTimestamp="2026-01-26 16:07:55 +0000 UTC" firstStartedPulling="2026-01-26 16:08:01.244709003 +0000 UTC m=+806.730437710" lastFinishedPulling="2026-01-26 16:08:11.763639048 +0000 UTC m=+817.249367755" observedRunningTime="2026-01-26 16:08:12.719929263 +0000 UTC m=+818.205657970" watchObservedRunningTime="2026-01-26 16:08:12.722387315 +0000 UTC m=+818.208116022" Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.739710 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-fq6ff" podStartSLOduration=4.739689032 podStartE2EDuration="4.739689032s" podCreationTimestamp="2026-01-26 16:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:08:12.7364699 +0000 UTC m=+818.222198617" watchObservedRunningTime="2026-01-26 16:08:12.739689032 +0000 UTC m=+818.225417739" Jan 26 16:08:12 crc kubenswrapper[4880]: I0126 16:08:12.789900 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" podStartSLOduration=8.237147959 podStartE2EDuration="18.789868758s" podCreationTimestamp="2026-01-26 16:07:54 +0000 UTC" firstStartedPulling="2026-01-26 16:08:01.244745034 +0000 UTC m=+806.730473741" lastFinishedPulling="2026-01-26 16:08:11.797465833 +0000 UTC m=+817.283194540" observedRunningTime="2026-01-26 16:08:12.782563514 +0000 UTC m=+818.268292221" watchObservedRunningTime="2026-01-26 16:08:12.789868758 +0000 UTC m=+818.275597465" Jan 26 16:08:14 crc kubenswrapper[4880]: I0126 16:08:14.471291 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-dwwh5" Jan 26 16:08:16 crc kubenswrapper[4880]: I0126 16:08:16.751146 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-sz2ch" event={"ID":"f77203c1-a597-4711-aac7-0f20f88aef3f","Type":"ContainerStarted","Data":"ffb7d7f5ac3003eebdbd20258fa18b56f16b6096ce833fcc0b1acfb5dd410db4"} Jan 26 16:08:16 crc kubenswrapper[4880]: I0126 16:08:16.765292 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-sz2ch" podStartSLOduration=5.248589346 podStartE2EDuration="8.76526853s" podCreationTimestamp="2026-01-26 16:08:08 +0000 UTC" firstStartedPulling="2026-01-26 16:08:12.165300535 +0000 UTC m=+817.651029242" lastFinishedPulling="2026-01-26 16:08:15.681979719 +0000 UTC m=+821.167708426" observedRunningTime="2026-01-26 16:08:16.763805233 +0000 UTC m=+822.249533940" watchObservedRunningTime="2026-01-26 16:08:16.76526853 +0000 UTC m=+822.250997237" Jan 26 16:08:16 crc kubenswrapper[4880]: I0126 16:08:16.896343 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="fff0a88a-97cf-45a6-8de4-4568e40bac22" containerName="elasticsearch" probeResult="failure" output=< Jan 26 16:08:16 crc kubenswrapper[4880]: {"timestamp": "2026-01-26T16:08:16+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 16:08:16 crc kubenswrapper[4880]: > Jan 26 16:08:19 crc kubenswrapper[4880]: I0126 16:08:19.166568 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:19 crc kubenswrapper[4880]: I0126 16:08:19.166972 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:19 crc kubenswrapper[4880]: I0126 16:08:19.206267 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:19 crc kubenswrapper[4880]: I0126 16:08:19.741510 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-v28tf" Jan 26 16:08:21 crc kubenswrapper[4880]: I0126 16:08:21.908987 4880 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="fff0a88a-97cf-45a6-8de4-4568e40bac22" containerName="elasticsearch" probeResult="failure" output=< Jan 26 16:08:21 crc kubenswrapper[4880]: {"timestamp": "2026-01-26T16:08:21+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 16:08:21 crc kubenswrapper[4880]: > Jan 26 16:08:27 crc kubenswrapper[4880]: I0126 16:08:27.131384 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 16:08:29 crc kubenswrapper[4880]: I0126 16:08:29.205448 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-sz2ch" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.081623 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5"] Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.084206 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.110263 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5"] Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.111342 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.111528 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.111786 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd2l6\" (UniqueName: \"kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.212270 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.212337 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.212398 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd2l6\" (UniqueName: \"kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.213317 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.213369 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.232758 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd2l6\" (UniqueName: \"kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6\") pod \"e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.410975 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.907629 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl"] Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.910754 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.916572 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.928754 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl"] Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.966905 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.967114 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zrc7\" (UniqueName: \"kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:31 crc kubenswrapper[4880]: I0126 16:08:31.967196 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.067938 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zrc7\" (UniqueName: \"kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.068341 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.068499 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.068996 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.069176 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.092019 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zrc7\" (UniqueName: \"kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.203264 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5"] Jan 26 16:08:32 crc kubenswrapper[4880]: W0126 16:08:32.208863 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd436acb0_594d_4a0e_afdf_d0810f7aec3e.slice/crio-812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f WatchSource:0}: Error finding container 812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f: Status 404 returned error can't find the container with id 812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.238547 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.491389 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl"] Jan 26 16:08:32 crc kubenswrapper[4880]: W0126 16:08:32.502905 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda039001a_3baa_48a4_9181_235e48958939.slice/crio-8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d WatchSource:0}: Error finding container 8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d: Status 404 returned error can't find the container with id 8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.865446 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx"] Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.867210 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.881089 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx"] Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.917298 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" event={"ID":"a039001a-3baa-48a4-9181-235e48958939","Type":"ContainerStarted","Data":"8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d"} Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.937158 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerStarted","Data":"41f42d02a6f439e9a6a14729b307ef0db41c86a711ec2c3d3eb067e995da941c"} Jan 26 16:08:32 crc kubenswrapper[4880]: I0126 16:08:32.937216 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerStarted","Data":"812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f"} Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.020621 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxtb8\" (UniqueName: \"kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.020741 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.020852 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.122358 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxtb8\" (UniqueName: \"kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.122548 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.122650 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.123194 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.123286 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.145327 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxtb8\" (UniqueName: \"kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8\") pod \"2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.184447 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.531228 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx"] Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.945820 4880 generic.go:334] "Generic (PLEG): container finished" podID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerID="edf56654d8a89c53b556300971e77da0608be5ee3b6e90571c910396fa362418" exitCode=0 Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.945922 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerDied","Data":"edf56654d8a89c53b556300971e77da0608be5ee3b6e90571c910396fa362418"} Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.945995 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerStarted","Data":"477bc67bff6a0821938c0892af99b0d8195422fd41ad78f73dab51bc877e5654"} Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.947804 4880 generic.go:334] "Generic (PLEG): container finished" podID="a039001a-3baa-48a4-9181-235e48958939" containerID="c33a6c5ddba0189d424e1c420a10e1df12dfece0253ce12971b8bcf86f445282" exitCode=0 Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.947855 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" event={"ID":"a039001a-3baa-48a4-9181-235e48958939","Type":"ContainerDied","Data":"c33a6c5ddba0189d424e1c420a10e1df12dfece0253ce12971b8bcf86f445282"} Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.951883 4880 generic.go:334] "Generic (PLEG): container finished" podID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerID="41f42d02a6f439e9a6a14729b307ef0db41c86a711ec2c3d3eb067e995da941c" exitCode=0 Jan 26 16:08:33 crc kubenswrapper[4880]: I0126 16:08:33.951916 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerDied","Data":"41f42d02a6f439e9a6a14729b307ef0db41c86a711ec2c3d3eb067e995da941c"} Jan 26 16:08:34 crc kubenswrapper[4880]: I0126 16:08:34.960279 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerStarted","Data":"8c2dcdbc863b331093753f706f78e2abf813b31af7bf284eb0be375efc136f95"} Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.047160 4880 generic.go:334] "Generic (PLEG): container finished" podID="a039001a-3baa-48a4-9181-235e48958939" containerID="80c5f8fbe97c6483753185f9d85e5fdb923726069ccb73fb2dfe31ee986642e3" exitCode=0 Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.047214 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" event={"ID":"a039001a-3baa-48a4-9181-235e48958939","Type":"ContainerDied","Data":"80c5f8fbe97c6483753185f9d85e5fdb923726069ccb73fb2dfe31ee986642e3"} Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.049538 4880 generic.go:334] "Generic (PLEG): container finished" podID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerID="782a22d8f033a341bd09f334e5b391c6d136557eeb635bdf9d36c92a51559c7f" exitCode=0 Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.049616 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerDied","Data":"782a22d8f033a341bd09f334e5b391c6d136557eeb635bdf9d36c92a51559c7f"} Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.067006 4880 generic.go:334] "Generic (PLEG): container finished" podID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerID="8c2dcdbc863b331093753f706f78e2abf813b31af7bf284eb0be375efc136f95" exitCode=0 Jan 26 16:08:36 crc kubenswrapper[4880]: I0126 16:08:36.067100 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerDied","Data":"8c2dcdbc863b331093753f706f78e2abf813b31af7bf284eb0be375efc136f95"} Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.075838 4880 generic.go:334] "Generic (PLEG): container finished" podID="a039001a-3baa-48a4-9181-235e48958939" containerID="2ef1b61fe5e2ba666a2210656477a116caef9fa2f6e364de8022c282dc06f430" exitCode=0 Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.076017 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" event={"ID":"a039001a-3baa-48a4-9181-235e48958939","Type":"ContainerDied","Data":"2ef1b61fe5e2ba666a2210656477a116caef9fa2f6e364de8022c282dc06f430"} Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.080232 4880 generic.go:334] "Generic (PLEG): container finished" podID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerID="e103ec23294867a87f11b82ba726948f61d2d98ef93f2e2c3df1a6279c4244dc" exitCode=0 Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.080386 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerDied","Data":"e103ec23294867a87f11b82ba726948f61d2d98ef93f2e2c3df1a6279c4244dc"} Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.082819 4880 generic.go:334] "Generic (PLEG): container finished" podID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerID="33556413a9d2008bccb3f50e991d18f82a965f3b832078eaaae112f098a1b467" exitCode=0 Jan 26 16:08:37 crc kubenswrapper[4880]: I0126 16:08:37.082855 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerDied","Data":"33556413a9d2008bccb3f50e991d18f82a965f3b832078eaaae112f098a1b467"} Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.490507 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.498660 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.518040 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.527015 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd2l6\" (UniqueName: \"kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6\") pod \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.527055 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxtb8\" (UniqueName: \"kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8\") pod \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.527077 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util\") pod \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.538651 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8" (OuterVolumeSpecName: "kube-api-access-lxtb8") pod "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" (UID: "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79"). InnerVolumeSpecName "kube-api-access-lxtb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.539638 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6" (OuterVolumeSpecName: "kube-api-access-vd2l6") pod "d436acb0-594d-4a0e-afdf-d0810f7aec3e" (UID: "d436acb0-594d-4a0e-afdf-d0810f7aec3e"). InnerVolumeSpecName "kube-api-access-vd2l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628077 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle\") pod \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\" (UID: \"d436acb0-594d-4a0e-afdf-d0810f7aec3e\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628126 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle\") pod \"a039001a-3baa-48a4-9181-235e48958939\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628199 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util\") pod \"a039001a-3baa-48a4-9181-235e48958939\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628239 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zrc7\" (UniqueName: \"kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7\") pod \"a039001a-3baa-48a4-9181-235e48958939\" (UID: \"a039001a-3baa-48a4-9181-235e48958939\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628266 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle\") pod \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628296 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util\") pod \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\" (UID: \"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79\") " Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628606 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxtb8\" (UniqueName: \"kubernetes.io/projected/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-kube-api-access-lxtb8\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628630 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd2l6\" (UniqueName: \"kubernetes.io/projected/d436acb0-594d-4a0e-afdf-d0810f7aec3e-kube-api-access-vd2l6\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628821 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle" (OuterVolumeSpecName: "bundle") pod "d436acb0-594d-4a0e-afdf-d0810f7aec3e" (UID: "d436acb0-594d-4a0e-afdf-d0810f7aec3e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.628982 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle" (OuterVolumeSpecName: "bundle") pod "a039001a-3baa-48a4-9181-235e48958939" (UID: "a039001a-3baa-48a4-9181-235e48958939"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.629028 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle" (OuterVolumeSpecName: "bundle") pod "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" (UID: "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.631955 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7" (OuterVolumeSpecName: "kube-api-access-7zrc7") pod "a039001a-3baa-48a4-9181-235e48958939" (UID: "a039001a-3baa-48a4-9181-235e48958939"). InnerVolumeSpecName "kube-api-access-7zrc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.643904 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util" (OuterVolumeSpecName: "util") pod "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" (UID: "a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.687815 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util" (OuterVolumeSpecName: "util") pod "a039001a-3baa-48a4-9181-235e48958939" (UID: "a039001a-3baa-48a4-9181-235e48958939"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.712484 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util" (OuterVolumeSpecName: "util") pod "d436acb0-594d-4a0e-afdf-d0810f7aec3e" (UID: "d436acb0-594d-4a0e-afdf-d0810f7aec3e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730084 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730142 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730157 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a039001a-3baa-48a4-9181-235e48958939-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730169 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zrc7\" (UniqueName: \"kubernetes.io/projected/a039001a-3baa-48a4-9181-235e48958939-kube-api-access-7zrc7\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730222 4880 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730234 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:38 crc kubenswrapper[4880]: I0126 16:08:38.730243 4880 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d436acb0-594d-4a0e-afdf-d0810f7aec3e-util\") on node \"crc\" DevicePath \"\"" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.100417 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.100383 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/2ed151d2bd3e125f2453293dd520598a3bcb66264bbe0a9abec30f3a5essnmx" event={"ID":"a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79","Type":"ContainerDied","Data":"477bc67bff6a0821938c0892af99b0d8195422fd41ad78f73dab51bc877e5654"} Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.101060 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="477bc67bff6a0821938c0892af99b0d8195422fd41ad78f73dab51bc877e5654" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.103834 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" event={"ID":"a039001a-3baa-48a4-9181-235e48958939","Type":"ContainerDied","Data":"8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d"} Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.104002 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a28196206ed2303aefadb4097f4065d088ff2e935e87124b9d09e773a13311d" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.103957 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.106856 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" event={"ID":"d436acb0-594d-4a0e-afdf-d0810f7aec3e","Type":"ContainerDied","Data":"812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f"} Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.106938 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="812dac95c80096aeade69e34705ac766d46a4eca1a27d956ec76710439241f1f" Jan 26 16:08:39 crc kubenswrapper[4880]: I0126 16:08:39.106883 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/e2ca482ba9d5013d52294c83a5aba8d96bfcab343aa05dec9349fd2215sqmq5" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.526768 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-zl26f"] Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527554 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527572 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527588 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527595 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527612 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527619 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527629 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527634 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527642 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527648 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527654 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527660 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527668 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527674 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527687 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527695 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="pull" Jan 26 16:08:45 crc kubenswrapper[4880]: E0126 16:08:45.527705 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527710 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="util" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527843 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="a039001a-3baa-48a4-9181-235e48958939" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527862 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3bfdf71-ae4f-4e68-9b86-fc85ab3deb79" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.527871 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="d436acb0-594d-4a0e-afdf-d0810f7aec3e" containerName="extract" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.528319 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.531387 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-r9qrr" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.543472 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-zl26f"] Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.659600 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dxjx\" (UniqueName: \"kubernetes.io/projected/2ec80ff5-216f-4dc3-bf68-2883d38563b9-kube-api-access-6dxjx\") pod \"interconnect-operator-5bb49f789d-zl26f\" (UID: \"2ec80ff5-216f-4dc3-bf68-2883d38563b9\") " pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.760614 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dxjx\" (UniqueName: \"kubernetes.io/projected/2ec80ff5-216f-4dc3-bf68-2883d38563b9-kube-api-access-6dxjx\") pod \"interconnect-operator-5bb49f789d-zl26f\" (UID: \"2ec80ff5-216f-4dc3-bf68-2883d38563b9\") " pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.780120 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dxjx\" (UniqueName: \"kubernetes.io/projected/2ec80ff5-216f-4dc3-bf68-2883d38563b9-kube-api-access-6dxjx\") pod \"interconnect-operator-5bb49f789d-zl26f\" (UID: \"2ec80ff5-216f-4dc3-bf68-2883d38563b9\") " pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" Jan 26 16:08:45 crc kubenswrapper[4880]: I0126 16:08:45.854294 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.163277 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-zl26f"] Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.776043 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw"] Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.777077 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.779403 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-sh26h" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.793790 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw"] Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.886144 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/695785ce-59d7-4136-83c3-70ad2ab08f8b-runner\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.886314 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwxp5\" (UniqueName: \"kubernetes.io/projected/695785ce-59d7-4136-83c3-70ad2ab08f8b-kube-api-access-zwxp5\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.987916 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwxp5\" (UniqueName: \"kubernetes.io/projected/695785ce-59d7-4136-83c3-70ad2ab08f8b-kube-api-access-zwxp5\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.987998 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/695785ce-59d7-4136-83c3-70ad2ab08f8b-runner\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:46 crc kubenswrapper[4880]: I0126 16:08:46.988544 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/695785ce-59d7-4136-83c3-70ad2ab08f8b-runner\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:47 crc kubenswrapper[4880]: I0126 16:08:47.008150 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwxp5\" (UniqueName: \"kubernetes.io/projected/695785ce-59d7-4136-83c3-70ad2ab08f8b-kube-api-access-zwxp5\") pod \"smart-gateway-operator-77c9d9f969-8jlvw\" (UID: \"695785ce-59d7-4136-83c3-70ad2ab08f8b\") " pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:47 crc kubenswrapper[4880]: I0126 16:08:47.096340 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" Jan 26 16:08:47 crc kubenswrapper[4880]: I0126 16:08:47.176842 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" event={"ID":"2ec80ff5-216f-4dc3-bf68-2883d38563b9","Type":"ContainerStarted","Data":"6a7adf8b808882b65580e520e9d366b7f93437210844351059bdb5f331261be5"} Jan 26 16:08:47 crc kubenswrapper[4880]: I0126 16:08:47.335497 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw"] Jan 26 16:08:47 crc kubenswrapper[4880]: W0126 16:08:47.344162 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695785ce_59d7_4136_83c3_70ad2ab08f8b.slice/crio-085c54c00a98e45ab17abdac9c40752b3814958cb1999e6ad6ef257d8ae74f1c WatchSource:0}: Error finding container 085c54c00a98e45ab17abdac9c40752b3814958cb1999e6ad6ef257d8ae74f1c: Status 404 returned error can't find the container with id 085c54c00a98e45ab17abdac9c40752b3814958cb1999e6ad6ef257d8ae74f1c Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.188221 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" event={"ID":"695785ce-59d7-4136-83c3-70ad2ab08f8b","Type":"ContainerStarted","Data":"085c54c00a98e45ab17abdac9c40752b3814958cb1999e6ad6ef257d8ae74f1c"} Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.581157 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-68688768b9-p2q6k"] Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.582417 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.586042 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-gsg59" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.596404 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-68688768b9-p2q6k"] Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.714607 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf5qm\" (UniqueName: \"kubernetes.io/projected/cf21b426-5786-4406-be9a-469a13a1551b-kube-api-access-lf5qm\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.715082 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/cf21b426-5786-4406-be9a-469a13a1551b-runner\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.816773 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf5qm\" (UniqueName: \"kubernetes.io/projected/cf21b426-5786-4406-be9a-469a13a1551b-kube-api-access-lf5qm\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.816924 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/cf21b426-5786-4406-be9a-469a13a1551b-runner\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.819924 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/cf21b426-5786-4406-be9a-469a13a1551b-runner\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.844065 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf5qm\" (UniqueName: \"kubernetes.io/projected/cf21b426-5786-4406-be9a-469a13a1551b-kube-api-access-lf5qm\") pod \"service-telemetry-operator-68688768b9-p2q6k\" (UID: \"cf21b426-5786-4406-be9a-469a13a1551b\") " pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:48 crc kubenswrapper[4880]: I0126 16:08:48.946655 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" Jan 26 16:08:49 crc kubenswrapper[4880]: I0126 16:08:49.702092 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-68688768b9-p2q6k"] Jan 26 16:08:49 crc kubenswrapper[4880]: W0126 16:08:49.726784 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf21b426_5786_4406_be9a_469a13a1551b.slice/crio-33746bffbd3669a14912f5f383e6c1ec0626d318fe05b76ad1d9170c59aa71f4 WatchSource:0}: Error finding container 33746bffbd3669a14912f5f383e6c1ec0626d318fe05b76ad1d9170c59aa71f4: Status 404 returned error can't find the container with id 33746bffbd3669a14912f5f383e6c1ec0626d318fe05b76ad1d9170c59aa71f4 Jan 26 16:08:50 crc kubenswrapper[4880]: I0126 16:08:50.421704 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" event={"ID":"cf21b426-5786-4406-be9a-469a13a1551b","Type":"ContainerStarted","Data":"33746bffbd3669a14912f5f383e6c1ec0626d318fe05b76ad1d9170c59aa71f4"} Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.786579 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:latest" Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.787494 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:quay.io/infrawatch/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1768085178,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zwxp5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-77c9d9f969-8jlvw_service-telemetry(695785ce-59d7-4136-83c3-70ad2ab08f8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.789149 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" podUID="695785ce-59d7-4136-83c3-70ad2ab08f8b" Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.928358 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/service-telemetry-operator:latest" Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.928789 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/service-telemetry-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:service-telemetry-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_WEBHOOK_SNMP_IMAGE,Value:quay.io/infrawatch/prometheus-webhook-snmp:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_IMAGE,Value:quay.io/prometheus/prometheus:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER_IMAGE,Value:quay.io/prometheus/alertmanager:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:service-telemetry-operator.v1.5.1768085182,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lf5qm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod service-telemetry-operator-68688768b9-p2q6k_service-telemetry(cf21b426-5786-4406-be9a-469a13a1551b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:09:14 crc kubenswrapper[4880]: E0126 16:09:14.930797 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" podUID="cf21b426-5786-4406-be9a-469a13a1551b" Jan 26 16:09:15 crc kubenswrapper[4880]: I0126 16:09:15.800108 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" event={"ID":"2ec80ff5-216f-4dc3-bf68-2883d38563b9","Type":"ContainerStarted","Data":"81d4b78f28798eb6863d407ce809a54d6ab3905e1d920b81368f5eb819a05272"} Jan 26 16:09:15 crc kubenswrapper[4880]: E0126 16:09:15.801127 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:latest\\\"\"" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" podUID="695785ce-59d7-4136-83c3-70ad2ab08f8b" Jan 26 16:09:15 crc kubenswrapper[4880]: E0126 16:09:15.802307 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/service-telemetry-operator:latest\\\"\"" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" podUID="cf21b426-5786-4406-be9a-469a13a1551b" Jan 26 16:09:15 crc kubenswrapper[4880]: I0126 16:09:15.842350 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-zl26f" podStartSLOduration=10.500438031 podStartE2EDuration="30.842303522s" podCreationTimestamp="2026-01-26 16:08:45 +0000 UTC" firstStartedPulling="2026-01-26 16:08:46.173623402 +0000 UTC m=+851.659352109" lastFinishedPulling="2026-01-26 16:09:06.515488883 +0000 UTC m=+872.001217600" observedRunningTime="2026-01-26 16:09:15.841316148 +0000 UTC m=+881.327044855" watchObservedRunningTime="2026-01-26 16:09:15.842303522 +0000 UTC m=+881.328032239" Jan 26 16:09:27 crc kubenswrapper[4880]: I0126 16:09:27.876940 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" event={"ID":"695785ce-59d7-4136-83c3-70ad2ab08f8b","Type":"ContainerStarted","Data":"62442377ccc69b243d3ff078a26c31f83dfcc1b0efd453d973b3ae300156fbcb"} Jan 26 16:09:27 crc kubenswrapper[4880]: I0126 16:09:27.903701 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-77c9d9f969-8jlvw" podStartSLOduration=2.1543646929999998 podStartE2EDuration="41.903671046s" podCreationTimestamp="2026-01-26 16:08:46 +0000 UTC" firstStartedPulling="2026-01-26 16:08:47.346036902 +0000 UTC m=+852.831765609" lastFinishedPulling="2026-01-26 16:09:27.095343255 +0000 UTC m=+892.581071962" observedRunningTime="2026-01-26 16:09:27.895015127 +0000 UTC m=+893.380743844" watchObservedRunningTime="2026-01-26 16:09:27.903671046 +0000 UTC m=+893.389399773" Jan 26 16:09:28 crc kubenswrapper[4880]: I0126 16:09:28.884386 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" event={"ID":"cf21b426-5786-4406-be9a-469a13a1551b","Type":"ContainerStarted","Data":"e5845d619e949935039e1c287c6a4f91e6d6e0b71e7bb2b553d917361e9dbd63"} Jan 26 16:09:28 crc kubenswrapper[4880]: I0126 16:09:28.904547 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-68688768b9-p2q6k" podStartSLOduration=2.722830714 podStartE2EDuration="40.904526905s" podCreationTimestamp="2026-01-26 16:08:48 +0000 UTC" firstStartedPulling="2026-01-26 16:08:49.730465769 +0000 UTC m=+855.216194476" lastFinishedPulling="2026-01-26 16:09:27.91216194 +0000 UTC m=+893.397890667" observedRunningTime="2026-01-26 16:09:28.899666723 +0000 UTC m=+894.385395440" watchObservedRunningTime="2026-01-26 16:09:28.904526905 +0000 UTC m=+894.390255612" Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.819430 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.821174 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.833484 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.958210 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.958281 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:34 crc kubenswrapper[4880]: I0126 16:09:34.958488 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp94m\" (UniqueName: \"kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.060108 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp94m\" (UniqueName: \"kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.060248 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.060288 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.060981 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.061004 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.097795 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp94m\" (UniqueName: \"kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m\") pod \"community-operators-5mhkv\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.139903 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.400159 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.927591 4880 generic.go:334] "Generic (PLEG): container finished" podID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerID="cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a" exitCode=0 Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.927758 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerDied","Data":"cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a"} Jan 26 16:09:35 crc kubenswrapper[4880]: I0126 16:09:35.927888 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerStarted","Data":"6c4d36e9f125bad3ef0838a2db3d6f8ae6cf5636290e15dd00dfa2132c4dd657"} Jan 26 16:09:36 crc kubenswrapper[4880]: I0126 16:09:36.936336 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerStarted","Data":"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12"} Jan 26 16:09:37 crc kubenswrapper[4880]: E0126 16:09:37.173053 4880 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb23e5c2_a43e_4584_b7dc_03d0a3e48b2b.slice/crio-conmon-71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12.scope\": RecentStats: unable to find data in memory cache]" Jan 26 16:09:37 crc kubenswrapper[4880]: I0126 16:09:37.944409 4880 generic.go:334] "Generic (PLEG): container finished" podID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerID="71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12" exitCode=0 Jan 26 16:09:37 crc kubenswrapper[4880]: I0126 16:09:37.944528 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerDied","Data":"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12"} Jan 26 16:09:38 crc kubenswrapper[4880]: I0126 16:09:38.962835 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerStarted","Data":"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef"} Jan 26 16:09:38 crc kubenswrapper[4880]: I0126 16:09:38.988094 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5mhkv" podStartSLOduration=2.495917994 podStartE2EDuration="4.988056201s" podCreationTimestamp="2026-01-26 16:09:34 +0000 UTC" firstStartedPulling="2026-01-26 16:09:35.929312056 +0000 UTC m=+901.415040763" lastFinishedPulling="2026-01-26 16:09:38.421450263 +0000 UTC m=+903.907178970" observedRunningTime="2026-01-26 16:09:38.981584829 +0000 UTC m=+904.467313556" watchObservedRunningTime="2026-01-26 16:09:38.988056201 +0000 UTC m=+904.473784908" Jan 26 16:09:40 crc kubenswrapper[4880]: I0126 16:09:40.700018 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:09:40 crc kubenswrapper[4880]: I0126 16:09:40.700187 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:09:45 crc kubenswrapper[4880]: I0126 16:09:45.140649 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:45 crc kubenswrapper[4880]: I0126 16:09:45.141142 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:45 crc kubenswrapper[4880]: I0126 16:09:45.186223 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:46 crc kubenswrapper[4880]: I0126 16:09:46.137323 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:47 crc kubenswrapper[4880]: I0126 16:09:47.012516 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:48 crc kubenswrapper[4880]: I0126 16:09:48.044017 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5mhkv" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="registry-server" containerID="cri-o://b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef" gracePeriod=2 Jan 26 16:09:48 crc kubenswrapper[4880]: I0126 16:09:48.950836 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.051913 4880 generic.go:334] "Generic (PLEG): container finished" podID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerID="b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef" exitCode=0 Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.051967 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerDied","Data":"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef"} Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.051998 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mhkv" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.052031 4880 scope.go:117] "RemoveContainer" containerID="b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.052004 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mhkv" event={"ID":"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b","Type":"ContainerDied","Data":"6c4d36e9f125bad3ef0838a2db3d6f8ae6cf5636290e15dd00dfa2132c4dd657"} Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.057208 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content\") pod \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.057267 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp94m\" (UniqueName: \"kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m\") pod \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.057423 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities\") pod \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\" (UID: \"bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b\") " Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.058636 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities" (OuterVolumeSpecName: "utilities") pod "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" (UID: "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.065678 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m" (OuterVolumeSpecName: "kube-api-access-zp94m") pod "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" (UID: "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b"). InnerVolumeSpecName "kube-api-access-zp94m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.077394 4880 scope.go:117] "RemoveContainer" containerID="71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.098765 4880 scope.go:117] "RemoveContainer" containerID="cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.119770 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" (UID: "bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.123883 4880 scope.go:117] "RemoveContainer" containerID="b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef" Jan 26 16:09:49 crc kubenswrapper[4880]: E0126 16:09:49.124655 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef\": container with ID starting with b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef not found: ID does not exist" containerID="b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.124722 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef"} err="failed to get container status \"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef\": rpc error: code = NotFound desc = could not find container \"b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef\": container with ID starting with b00251eadd3bd9ca84bac5ef10c84f40722a5817510ec8a3923e1a0181eeb0ef not found: ID does not exist" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.124761 4880 scope.go:117] "RemoveContainer" containerID="71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12" Jan 26 16:09:49 crc kubenswrapper[4880]: E0126 16:09:49.125965 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12\": container with ID starting with 71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12 not found: ID does not exist" containerID="71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.125995 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12"} err="failed to get container status \"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12\": rpc error: code = NotFound desc = could not find container \"71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12\": container with ID starting with 71ed1fa3118cecb13cac9126ca210ff5d6da868a99e9ae2fb4fe0369b35f2a12 not found: ID does not exist" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.126014 4880 scope.go:117] "RemoveContainer" containerID="cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a" Jan 26 16:09:49 crc kubenswrapper[4880]: E0126 16:09:49.126411 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a\": container with ID starting with cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a not found: ID does not exist" containerID="cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.126452 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a"} err="failed to get container status \"cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a\": rpc error: code = NotFound desc = could not find container \"cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a\": container with ID starting with cab5dac13b5a89cbc7bc8ac8c544abf09f515b6454e1bb14f4f281afa767c79a not found: ID does not exist" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.158761 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp94m\" (UniqueName: \"kubernetes.io/projected/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-kube-api-access-zp94m\") on node \"crc\" DevicePath \"\"" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.158804 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.158818 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.389741 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:49 crc kubenswrapper[4880]: I0126 16:09:49.394324 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5mhkv"] Jan 26 16:09:50 crc kubenswrapper[4880]: I0126 16:09:50.231719 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" path="/var/lib/kubelet/pods/bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b/volumes" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.077391 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:09:55 crc kubenswrapper[4880]: E0126 16:09:55.078118 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="extract-utilities" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.078146 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="extract-utilities" Jan 26 16:09:55 crc kubenswrapper[4880]: E0126 16:09:55.078159 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="extract-content" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.078217 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="extract-content" Jan 26 16:09:55 crc kubenswrapper[4880]: E0126 16:09:55.078246 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="registry-server" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.078256 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="registry-server" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.078453 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb23e5c2-a43e-4584-b7dc-03d0a3e48b2b" containerName="registry-server" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.079160 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.080899 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.081257 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.081927 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-pbkdw" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.082142 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.083595 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.086262 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.086276 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.105359 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141225 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141293 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141365 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141394 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141514 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141547 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq8fm\" (UniqueName: \"kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.141608 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242555 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242756 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242856 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242889 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq8fm\" (UniqueName: \"kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242917 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242955 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.242989 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.244059 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.249620 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.249730 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.250226 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.258540 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.271402 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.285137 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq8fm\" (UniqueName: \"kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm\") pod \"default-interconnect-68864d46cb-v69cb\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.399634 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:09:55 crc kubenswrapper[4880]: I0126 16:09:55.727718 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:09:56 crc kubenswrapper[4880]: I0126 16:09:56.103654 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" event={"ID":"6bfefc28-b018-4cd5-b468-cd6f4fede198","Type":"ContainerStarted","Data":"471c0a57da0df6462858939ff7e767ffb51b3915f8893a27364afd7b10794d1e"} Jan 26 16:10:02 crc kubenswrapper[4880]: I0126 16:10:02.149671 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" event={"ID":"6bfefc28-b018-4cd5-b468-cd6f4fede198","Type":"ContainerStarted","Data":"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a"} Jan 26 16:10:02 crc kubenswrapper[4880]: I0126 16:10:02.175312 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" podStartSLOduration=1.820782079 podStartE2EDuration="7.175275627s" podCreationTimestamp="2026-01-26 16:09:55 +0000 UTC" firstStartedPulling="2026-01-26 16:09:55.760918175 +0000 UTC m=+921.246646882" lastFinishedPulling="2026-01-26 16:10:01.115411723 +0000 UTC m=+926.601140430" observedRunningTime="2026-01-26 16:10:02.1705535 +0000 UTC m=+927.656282207" watchObservedRunningTime="2026-01-26 16:10:02.175275627 +0000 UTC m=+927.661004334" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.795216 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.799914 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.806706 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.807545 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.807877 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.808002 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.807935 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.809921 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-mp4xn" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.810059 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.809958 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.809973 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.809974 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.829699 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998334 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998390 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998470 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-web-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998489 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998511 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/77ec4615-159f-4075-84f0-4d2515636f1f-config-out\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998533 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998559 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998583 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998621 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-tls-assets\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998641 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998687 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m724n\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-kube-api-access-m724n\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:05 crc kubenswrapper[4880]: I0126 16:10:05.998727 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101506 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-tls-assets\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101579 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101603 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m724n\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-kube-api-access-m724n\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101630 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101665 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101688 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101720 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-web-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101736 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101757 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/77ec4615-159f-4075-84f0-4d2515636f1f-config-out\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101775 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101800 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.101822 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.102557 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.103140 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: E0126 16:10:06.103505 4880 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 26 16:10:06 crc kubenswrapper[4880]: E0126 16:10:06.103675 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls podName:77ec4615-159f-4075-84f0-4d2515636f1f nodeName:}" failed. No retries permitted until 2026-01-26 16:10:06.603622941 +0000 UTC m=+932.089351678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "77ec4615-159f-4075-84f0-4d2515636f1f") : secret "default-prometheus-proxy-tls" not found Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.104161 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.104381 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/77ec4615-159f-4075-84f0-4d2515636f1f-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.106974 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/77ec4615-159f-4075-84f0-4d2515636f1f-config-out\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.107578 4880 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.107620 4880 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ec1feb7b3d152ef38598616321e736071c69b6bf6e71732ef3873baf647b5f70/globalmount\"" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.107638 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-tls-assets\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.110013 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-web-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.110822 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.117748 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-config\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.129848 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m724n\" (UniqueName: \"kubernetes.io/projected/77ec4615-159f-4075-84f0-4d2515636f1f-kube-api-access-m724n\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.143791 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ce4df020-e360-42bf-abc3-0b55811b4b9d\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: I0126 16:10:06.608624 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:06 crc kubenswrapper[4880]: E0126 16:10:06.608805 4880 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Jan 26 16:10:06 crc kubenswrapper[4880]: E0126 16:10:06.609095 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls podName:77ec4615-159f-4075-84f0-4d2515636f1f nodeName:}" failed. No retries permitted until 2026-01-26 16:10:07.6090748 +0000 UTC m=+933.094803507 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "77ec4615-159f-4075-84f0-4d2515636f1f") : secret "default-prometheus-proxy-tls" not found Jan 26 16:10:07 crc kubenswrapper[4880]: I0126 16:10:07.623068 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:07 crc kubenswrapper[4880]: I0126 16:10:07.628222 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/77ec4615-159f-4075-84f0-4d2515636f1f-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"77ec4615-159f-4075-84f0-4d2515636f1f\") " pod="service-telemetry/prometheus-default-0" Jan 26 16:10:07 crc kubenswrapper[4880]: I0126 16:10:07.628910 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:08 crc kubenswrapper[4880]: I0126 16:10:08.081610 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Jan 26 16:10:08 crc kubenswrapper[4880]: I0126 16:10:08.190516 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerStarted","Data":"ea5548f58d02bb4dfa1dfabb7919237a7dc012c7d90de9f7329d937cdada761c"} Jan 26 16:10:10 crc kubenswrapper[4880]: I0126 16:10:10.700342 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:10:10 crc kubenswrapper[4880]: I0126 16:10:10.700725 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:10:12 crc kubenswrapper[4880]: I0126 16:10:12.223142 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerStarted","Data":"69e6ba552fc5bb6d7e7e38ac4878c569e4257d7b7056ca1971b4b152656602d5"} Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.509168 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8"] Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.510472 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.518807 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8"] Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.672000 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frsgb\" (UniqueName: \"kubernetes.io/projected/edc9710b-d4f1-4a5a-b434-2e471dc1a491-kube-api-access-frsgb\") pod \"default-snmp-webhook-78bcbbdcff-j5sr8\" (UID: \"edc9710b-d4f1-4a5a-b434-2e471dc1a491\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.773832 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frsgb\" (UniqueName: \"kubernetes.io/projected/edc9710b-d4f1-4a5a-b434-2e471dc1a491-kube-api-access-frsgb\") pod \"default-snmp-webhook-78bcbbdcff-j5sr8\" (UID: \"edc9710b-d4f1-4a5a-b434-2e471dc1a491\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.810993 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frsgb\" (UniqueName: \"kubernetes.io/projected/edc9710b-d4f1-4a5a-b434-2e471dc1a491-kube-api-access-frsgb\") pod \"default-snmp-webhook-78bcbbdcff-j5sr8\" (UID: \"edc9710b-d4f1-4a5a-b434-2e471dc1a491\") " pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" Jan 26 16:10:16 crc kubenswrapper[4880]: I0126 16:10:16.848716 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" Jan 26 16:10:17 crc kubenswrapper[4880]: I0126 16:10:17.260284 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8"] Jan 26 16:10:18 crc kubenswrapper[4880]: I0126 16:10:18.269729 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" event={"ID":"edc9710b-d4f1-4a5a-b434-2e471dc1a491","Type":"ContainerStarted","Data":"c7507bd35094cee7bfb7c37f02bc3f1aedb3dd33884362fc349275ad644ad3f4"} Jan 26 16:10:19 crc kubenswrapper[4880]: I0126 16:10:19.284149 4880 generic.go:334] "Generic (PLEG): container finished" podID="77ec4615-159f-4075-84f0-4d2515636f1f" containerID="69e6ba552fc5bb6d7e7e38ac4878c569e4257d7b7056ca1971b4b152656602d5" exitCode=0 Jan 26 16:10:19 crc kubenswrapper[4880]: I0126 16:10:19.284194 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerDied","Data":"69e6ba552fc5bb6d7e7e38ac4878c569e4257d7b7056ca1971b4b152656602d5"} Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.005071 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.016897 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.149972 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.150053 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.150285 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-jcx4b" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.171716 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.172096 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.172688 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.186281 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.240715 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ptpx\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-kube-api-access-4ptpx\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.240992 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241108 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241192 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-config-volume\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241399 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a784e60c-653f-4125-8162-ec08686dee59-config-out\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241521 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241615 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-web-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241688 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.241796 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467235 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467287 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-config-volume\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467315 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a784e60c-653f-4125-8162-ec08686dee59-config-out\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467349 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467366 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-web-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467384 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467506 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467538 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ptpx\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-kube-api-access-4ptpx\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.467570 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: E0126 16:10:20.469182 4880 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:20 crc kubenswrapper[4880]: E0126 16:10:20.469482 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls podName:a784e60c-653f-4125-8162-ec08686dee59 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:20.96924198 +0000 UTC m=+946.454970687 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a784e60c-653f-4125-8162-ec08686dee59") : secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.475244 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-web-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.475712 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.477780 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a784e60c-653f-4125-8162-ec08686dee59-config-out\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.477928 4880 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.477976 4880 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7fbc05c74a9b37a8f3b12a11c1fe8a01759984e7e2f587032a0858c759719065/globalmount\"" pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.479277 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.481377 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.482583 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-config-volume\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.500423 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ptpx\" (UniqueName: \"kubernetes.io/projected/a784e60c-653f-4125-8162-ec08686dee59-kube-api-access-4ptpx\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:20 crc kubenswrapper[4880]: I0126 16:10:20.508382 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-594d57df-5a33-44b0-a4d5-207986f9ccf6\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:21 crc kubenswrapper[4880]: I0126 16:10:21.400247 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:21 crc kubenswrapper[4880]: E0126 16:10:21.400490 4880 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:21 crc kubenswrapper[4880]: E0126 16:10:21.400549 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls podName:a784e60c-653f-4125-8162-ec08686dee59 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:22.400529429 +0000 UTC m=+947.886258136 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a784e60c-653f-4125-8162-ec08686dee59") : secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:22 crc kubenswrapper[4880]: I0126 16:10:22.476997 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:22 crc kubenswrapper[4880]: E0126 16:10:22.477167 4880 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:22 crc kubenswrapper[4880]: E0126 16:10:22.477544 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls podName:a784e60c-653f-4125-8162-ec08686dee59 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:24.477523068 +0000 UTC m=+949.963251775 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a784e60c-653f-4125-8162-ec08686dee59") : secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:24 crc kubenswrapper[4880]: I0126 16:10:24.561356 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:24 crc kubenswrapper[4880]: E0126 16:10:24.561616 4880 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:24 crc kubenswrapper[4880]: E0126 16:10:24.561902 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls podName:a784e60c-653f-4125-8162-ec08686dee59 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:28.561877947 +0000 UTC m=+954.047606674 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a784e60c-653f-4125-8162-ec08686dee59") : secret "default-alertmanager-proxy-tls" not found Jan 26 16:10:28 crc kubenswrapper[4880]: I0126 16:10:28.577202 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:28 crc kubenswrapper[4880]: I0126 16:10:28.583562 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a784e60c-653f-4125-8162-ec08686dee59-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a784e60c-653f-4125-8162-ec08686dee59\") " pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:28 crc kubenswrapper[4880]: I0126 16:10:28.584126 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Jan 26 16:10:29 crc kubenswrapper[4880]: I0126 16:10:29.178070 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Jan 26 16:10:29 crc kubenswrapper[4880]: I0126 16:10:29.572516 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerStarted","Data":"61e875b7ec1b15c77c16fe27e3021606fecb115e2aa297c400ecef88a3dbf2cc"} Jan 26 16:10:29 crc kubenswrapper[4880]: I0126 16:10:29.575629 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" event={"ID":"edc9710b-d4f1-4a5a-b434-2e471dc1a491","Type":"ContainerStarted","Data":"d816592a0b36b53ae00b7f7ee4a9440c97f3bf46dc82bfaa4b91f5c9597e0195"} Jan 26 16:10:29 crc kubenswrapper[4880]: I0126 16:10:29.596597 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-78bcbbdcff-j5sr8" podStartSLOduration=2.296396535 podStartE2EDuration="13.59657002s" podCreationTimestamp="2026-01-26 16:10:16 +0000 UTC" firstStartedPulling="2026-01-26 16:10:17.272298009 +0000 UTC m=+942.758026706" lastFinishedPulling="2026-01-26 16:10:28.572471484 +0000 UTC m=+954.058200191" observedRunningTime="2026-01-26 16:10:29.593448601 +0000 UTC m=+955.079177308" watchObservedRunningTime="2026-01-26 16:10:29.59657002 +0000 UTC m=+955.082298727" Jan 26 16:10:31 crc kubenswrapper[4880]: I0126 16:10:31.594841 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerStarted","Data":"2453d47460abf7def7937e25f1c6e703ac5740157972eba6f5780ffd22acb442"} Jan 26 16:10:34 crc kubenswrapper[4880]: I0126 16:10:34.618934 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerStarted","Data":"561c22de25b01636908d1d94eb7cb70b8d60e0ed02b898e4dc24f0b57d716626"} Jan 26 16:10:36 crc kubenswrapper[4880]: I0126 16:10:36.633844 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerStarted","Data":"47273ce8072d16e6a0583ad7f400842bfbde6c0eba67374479b1e3458357e06c"} Jan 26 16:10:38 crc kubenswrapper[4880]: I0126 16:10:38.681918 4880 generic.go:334] "Generic (PLEG): container finished" podID="a784e60c-653f-4125-8162-ec08686dee59" containerID="2453d47460abf7def7937e25f1c6e703ac5740157972eba6f5780ffd22acb442" exitCode=0 Jan 26 16:10:38 crc kubenswrapper[4880]: I0126 16:10:38.682251 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerDied","Data":"2453d47460abf7def7937e25f1c6e703ac5740157972eba6f5780ffd22acb442"} Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.086216 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt"] Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.094236 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.097548 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.097838 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.098167 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-km9zf" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.102975 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt"] Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.104529 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.248879 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.248957 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.249022 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxsbf\" (UniqueName: \"kubernetes.io/projected/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-kube-api-access-xxsbf\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.249064 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.249131 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.350553 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.350712 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.350735 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.350799 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxsbf\" (UniqueName: \"kubernetes.io/projected/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-kube-api-access-xxsbf\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.350860 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: E0126 16:10:39.351660 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 26 16:10:39 crc kubenswrapper[4880]: E0126 16:10:39.351742 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls podName:3e64bef9-c93b-49e3-93ee-4161a2a7c89c nodeName:}" failed. No retries permitted until 2026-01-26 16:10:39.851721107 +0000 UTC m=+965.337449884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" (UID: "3e64bef9-c93b-49e3-93ee-4161a2a7c89c") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.352615 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.353145 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.364728 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.368771 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxsbf\" (UniqueName: \"kubernetes.io/projected/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-kube-api-access-xxsbf\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: I0126 16:10:39.857988 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:39 crc kubenswrapper[4880]: E0126 16:10:39.858294 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Jan 26 16:10:39 crc kubenswrapper[4880]: E0126 16:10:39.858372 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls podName:3e64bef9-c93b-49e3-93ee-4161a2a7c89c nodeName:}" failed. No retries permitted until 2026-01-26 16:10:40.858352835 +0000 UTC m=+966.344081532 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" (UID: "3e64bef9-c93b-49e3-93ee-4161a2a7c89c") : secret "default-cloud1-coll-meter-proxy-tls" not found Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.699582 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.699661 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.699714 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.700406 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.700505 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56" gracePeriod=600 Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.872055 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.876185 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e64bef9-c93b-49e3-93ee-4161a2a7c89c-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt\" (UID: \"3e64bef9-c93b-49e3-93ee-4161a2a7c89c\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:40 crc kubenswrapper[4880]: I0126 16:10:40.918057 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" Jan 26 16:10:41 crc kubenswrapper[4880]: I0126 16:10:41.709062 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56" exitCode=0 Jan 26 16:10:41 crc kubenswrapper[4880]: I0126 16:10:41.709104 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56"} Jan 26 16:10:41 crc kubenswrapper[4880]: I0126 16:10:41.709280 4880 scope.go:117] "RemoveContainer" containerID="9952a68c347bc7b5c1d0589c50d162cb14c6d8784c2f475f718b6909569962da" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.006412 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh"] Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.020583 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.024031 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh"] Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.029075 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.034046 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.094522 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.094600 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/04efd524-476a-4803-8996-9b16b3214c32-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.094670 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/04efd524-476a-4803-8996-9b16b3214c32-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.094749 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nshk8\" (UniqueName: \"kubernetes.io/projected/04efd524-476a-4803-8996-9b16b3214c32-kube-api-access-nshk8\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.094804 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.196895 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nshk8\" (UniqueName: \"kubernetes.io/projected/04efd524-476a-4803-8996-9b16b3214c32-kube-api-access-nshk8\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.197656 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: E0126 16:10:42.197749 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.197816 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.197857 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/04efd524-476a-4803-8996-9b16b3214c32-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.197962 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/04efd524-476a-4803-8996-9b16b3214c32-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: E0126 16:10:42.197982 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls podName:04efd524-476a-4803-8996-9b16b3214c32 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:42.697958945 +0000 UTC m=+968.183687652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" (UID: "04efd524-476a-4803-8996-9b16b3214c32") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.199262 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/04efd524-476a-4803-8996-9b16b3214c32-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.201529 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/04efd524-476a-4803-8996-9b16b3214c32-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.211503 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.216245 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nshk8\" (UniqueName: \"kubernetes.io/projected/04efd524-476a-4803-8996-9b16b3214c32-kube-api-access-nshk8\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: I0126 16:10:42.705178 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:42 crc kubenswrapper[4880]: E0126 16:10:42.705407 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 26 16:10:42 crc kubenswrapper[4880]: E0126 16:10:42.705500 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls podName:04efd524-476a-4803-8996-9b16b3214c32 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:43.705478374 +0000 UTC m=+969.191207081 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" (UID: "04efd524-476a-4803-8996-9b16b3214c32") : secret "default-cloud1-ceil-meter-proxy-tls" not found Jan 26 16:10:43 crc kubenswrapper[4880]: I0126 16:10:43.719856 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:43 crc kubenswrapper[4880]: I0126 16:10:43.726011 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/04efd524-476a-4803-8996-9b16b3214c32-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh\" (UID: \"04efd524-476a-4803-8996-9b16b3214c32\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:43 crc kubenswrapper[4880]: I0126 16:10:43.838123 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.334352 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt"] Jan 26 16:10:44 crc kubenswrapper[4880]: W0126 16:10:44.339074 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e64bef9_c93b_49e3_93ee_4161a2a7c89c.slice/crio-4e35e893d522304e46044fe7b23b28600a20f4eb6c8ad674cb082011289e668e WatchSource:0}: Error finding container 4e35e893d522304e46044fe7b23b28600a20f4eb6c8ad674cb082011289e668e: Status 404 returned error can't find the container with id 4e35e893d522304e46044fe7b23b28600a20f4eb6c8ad674cb082011289e668e Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.432278 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh"] Jan 26 16:10:44 crc kubenswrapper[4880]: W0126 16:10:44.436021 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04efd524_476a_4803_8996_9b16b3214c32.slice/crio-595b289739944847bc5ba7f23da31ed6cb335ec0beca1b307b934e3cc1c90bce WatchSource:0}: Error finding container 595b289739944847bc5ba7f23da31ed6cb335ec0beca1b307b934e3cc1c90bce: Status 404 returned error can't find the container with id 595b289739944847bc5ba7f23da31ed6cb335ec0beca1b307b934e3cc1c90bce Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.748045 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a"} Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.752648 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"595b289739944847bc5ba7f23da31ed6cb335ec0beca1b307b934e3cc1c90bce"} Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.754846 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"4e35e893d522304e46044fe7b23b28600a20f4eb6c8ad674cb082011289e668e"} Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.758192 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"77ec4615-159f-4075-84f0-4d2515636f1f","Type":"ContainerStarted","Data":"d608541050c30437d2acf6cedd245671b652994794f9f85ac28ded0c847136d2"} Jan 26 16:10:44 crc kubenswrapper[4880]: I0126 16:10:44.798724 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.63149427 podStartE2EDuration="40.798701643s" podCreationTimestamp="2026-01-26 16:10:04 +0000 UTC" firstStartedPulling="2026-01-26 16:10:08.09502275 +0000 UTC m=+933.580751457" lastFinishedPulling="2026-01-26 16:10:44.262230123 +0000 UTC m=+969.747958830" observedRunningTime="2026-01-26 16:10:44.793321409 +0000 UTC m=+970.279050116" watchObservedRunningTime="2026-01-26 16:10:44.798701643 +0000 UTC m=+970.284430350" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.060217 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4"] Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.061603 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.064197 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.064249 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.076835 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4"] Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.167836 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.167890 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.168037 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.168106 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.168330 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmmmc\" (UniqueName: \"kubernetes.io/projected/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-kube-api-access-dmmmc\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.269903 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmmmc\" (UniqueName: \"kubernetes.io/projected/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-kube-api-access-dmmmc\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.270010 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.270032 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.270054 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.270073 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.270887 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: E0126 16:10:46.272777 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 26 16:10:46 crc kubenswrapper[4880]: E0126 16:10:46.272907 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls podName:5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:46.77288009 +0000 UTC m=+972.258608877 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" (UID: "5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.273019 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.285331 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.289665 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmmmc\" (UniqueName: \"kubernetes.io/projected/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-kube-api-access-dmmmc\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.777751 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:46 crc kubenswrapper[4880]: E0126 16:10:46.777942 4880 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Jan 26 16:10:46 crc kubenswrapper[4880]: E0126 16:10:46.778380 4880 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls podName:5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16 nodeName:}" failed. No retries permitted until 2026-01-26 16:10:47.778359689 +0000 UTC m=+973.264088396 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" (UID: "5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16") : secret "default-cloud1-sens-meter-proxy-tls" not found Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.784467 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"01463541a71e56069c09c07a9984980f3e9d72a82f5b81f49b1931b89a236b69"} Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.787792 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerStarted","Data":"c480671731a084a760919e3dd940d943f974afd62f4ee1666445c902154a0466"} Jan 26 16:10:46 crc kubenswrapper[4880]: I0126 16:10:46.789146 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"61e83c5f1b7462d6948f9a1e05239e963b254e98d11eb2aad2315c41f81d91d4"} Jan 26 16:10:47 crc kubenswrapper[4880]: I0126 16:10:47.629934 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:47 crc kubenswrapper[4880]: I0126 16:10:47.793495 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:47 crc kubenswrapper[4880]: I0126 16:10:47.880738 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4\" (UID: \"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:47 crc kubenswrapper[4880]: I0126 16:10:47.886249 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" Jan 26 16:10:48 crc kubenswrapper[4880]: I0126 16:10:48.387033 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4"] Jan 26 16:10:48 crc kubenswrapper[4880]: W0126 16:10:48.401389 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5edcc1e8_51ca_4cc8_9de1_c3966d7bdb16.slice/crio-051a69c99bead6a21e82bc8c90bb0c73cf93b226fc62c41fd8f26ba319f5f733 WatchSource:0}: Error finding container 051a69c99bead6a21e82bc8c90bb0c73cf93b226fc62c41fd8f26ba319f5f733: Status 404 returned error can't find the container with id 051a69c99bead6a21e82bc8c90bb0c73cf93b226fc62c41fd8f26ba319f5f733 Jan 26 16:10:48 crc kubenswrapper[4880]: I0126 16:10:48.807965 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"051a69c99bead6a21e82bc8c90bb0c73cf93b226fc62c41fd8f26ba319f5f733"} Jan 26 16:10:49 crc kubenswrapper[4880]: I0126 16:10:49.818760 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerStarted","Data":"1be7323d5eece7ca6636667d419bc4f5b90553c0330c71edb82c3534c50f3917"} Jan 26 16:10:49 crc kubenswrapper[4880]: I0126 16:10:49.819339 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a784e60c-653f-4125-8162-ec08686dee59","Type":"ContainerStarted","Data":"3a6765d45d931020d32799b5af358629e2dbc9c1e31387b63a3233fb02786cea"} Jan 26 16:10:49 crc kubenswrapper[4880]: I0126 16:10:49.824131 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"9e64c05eb4c09b7d21eeeee5ca683a1fa58e74e170247413eeb5e3c01bae5e8f"} Jan 26 16:10:49 crc kubenswrapper[4880]: I0126 16:10:49.861194 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=20.988584473 podStartE2EDuration="31.861168645s" podCreationTimestamp="2026-01-26 16:10:18 +0000 UTC" firstStartedPulling="2026-01-26 16:10:38.684521579 +0000 UTC m=+964.170250286" lastFinishedPulling="2026-01-26 16:10:49.557105751 +0000 UTC m=+975.042834458" observedRunningTime="2026-01-26 16:10:49.847635499 +0000 UTC m=+975.333364216" watchObservedRunningTime="2026-01-26 16:10:49.861168645 +0000 UTC m=+975.346897352" Jan 26 16:10:52 crc kubenswrapper[4880]: I0126 16:10:52.629416 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:52 crc kubenswrapper[4880]: I0126 16:10:52.668684 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:52 crc kubenswrapper[4880]: I0126 16:10:52.884034 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.670618 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr"] Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.672360 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.674825 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.681960 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.685230 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr"] Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.851699 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/71cb321c-ed35-4d06-a487-eacea2dee247-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.851770 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/71cb321c-ed35-4d06-a487-eacea2dee247-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.851794 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/71cb321c-ed35-4d06-a487-eacea2dee247-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.851853 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pbhv\" (UniqueName: \"kubernetes.io/projected/71cb321c-ed35-4d06-a487-eacea2dee247-kube-api-access-4pbhv\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.954521 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/71cb321c-ed35-4d06-a487-eacea2dee247-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.954594 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/71cb321c-ed35-4d06-a487-eacea2dee247-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.954631 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/71cb321c-ed35-4d06-a487-eacea2dee247-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.954674 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pbhv\" (UniqueName: \"kubernetes.io/projected/71cb321c-ed35-4d06-a487-eacea2dee247-kube-api-access-4pbhv\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.956394 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/71cb321c-ed35-4d06-a487-eacea2dee247-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.957710 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/71cb321c-ed35-4d06-a487-eacea2dee247-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.963719 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/71cb321c-ed35-4d06-a487-eacea2dee247-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.970400 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pbhv\" (UniqueName: \"kubernetes.io/projected/71cb321c-ed35-4d06-a487-eacea2dee247-kube-api-access-4pbhv\") pod \"default-cloud1-coll-event-smartgateway-76756967c5-bpljr\" (UID: \"71cb321c-ed35-4d06-a487-eacea2dee247\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:55 crc kubenswrapper[4880]: I0126 16:10:55.995135 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" Jan 26 16:10:56 crc kubenswrapper[4880]: I0126 16:10:56.971896 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd"] Jan 26 16:10:56 crc kubenswrapper[4880]: I0126 16:10:56.973552 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:56 crc kubenswrapper[4880]: I0126 16:10:56.976373 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Jan 26 16:10:56 crc kubenswrapper[4880]: I0126 16:10:56.983499 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd"] Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.077592 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e07f674d-87fb-46ce-98bc-34c04cf788da-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.077637 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e07f674d-87fb-46ce-98bc-34c04cf788da-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.077670 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt6l7\" (UniqueName: \"kubernetes.io/projected/e07f674d-87fb-46ce-98bc-34c04cf788da-kube-api-access-tt6l7\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.077710 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e07f674d-87fb-46ce-98bc-34c04cf788da-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.179719 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e07f674d-87fb-46ce-98bc-34c04cf788da-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.179827 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e07f674d-87fb-46ce-98bc-34c04cf788da-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.179885 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt6l7\" (UniqueName: \"kubernetes.io/projected/e07f674d-87fb-46ce-98bc-34c04cf788da-kube-api-access-tt6l7\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.179933 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e07f674d-87fb-46ce-98bc-34c04cf788da-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.180654 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e07f674d-87fb-46ce-98bc-34c04cf788da-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.182183 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e07f674d-87fb-46ce-98bc-34c04cf788da-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.185833 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/e07f674d-87fb-46ce-98bc-34c04cf788da-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.200150 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt6l7\" (UniqueName: \"kubernetes.io/projected/e07f674d-87fb-46ce-98bc-34c04cf788da-kube-api-access-tt6l7\") pod \"default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd\" (UID: \"e07f674d-87fb-46ce-98bc-34c04cf788da\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:10:57 crc kubenswrapper[4880]: I0126 16:10:57.305365 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.531238 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/sg-bridge:latest" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.532301 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:quay.io/infrawatch/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/collectd/cloud1-telemetry --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxsbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt_service-telemetry(3e64bef9-c93b-49e3-93ee-4161a2a7c89c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.696610 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/sg-bridge:latest" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.696804 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:quay.io/infrawatch/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/anycast/ceilometer/cloud1-metering.sample --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 16384],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nshk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh_service-telemetry(04efd524-476a-4803-8996-9b16b3214c32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.707789 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/sg-bridge:latest" Jan 26 16:11:03 crc kubenswrapper[4880]: E0126 16:11:03.707979 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:bridge,Image:quay.io/infrawatch/sg-bridge:latest,Command:[],Args:[--amqp_url amqp://default-interconnect.service-telemetry.svc.cluster.local:5673/sensubility/cloud1-telemetry --block --stat_period 60 --rbc 15000 --count 0 --gw_unix /tmp/smartgateway --rbs 65535],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:socket-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmmmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4_service-telemetry(5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:11:04 crc kubenswrapper[4880]: I0126 16:11:04.132226 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr"] Jan 26 16:11:04 crc kubenswrapper[4880]: I0126 16:11:04.159979 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd"] Jan 26 16:11:04 crc kubenswrapper[4880]: W0126 16:11:04.167708 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode07f674d_87fb_46ce_98bc_34c04cf788da.slice/crio-4c8989c2304d3488c07c9cba84a112ba6b5622714014ca2ac091527d1e158143 WatchSource:0}: Error finding container 4c8989c2304d3488c07c9cba84a112ba6b5622714014ca2ac091527d1e158143: Status 404 returned error can't find the container with id 4c8989c2304d3488c07c9cba84a112ba6b5622714014ca2ac091527d1e158143 Jan 26 16:11:05 crc kubenswrapper[4880]: I0126 16:11:05.303107 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerStarted","Data":"77f346a0c8b332ad3f8151067c7f99cb989390598ab4b6d90dafea766aed5fe2"} Jan 26 16:11:05 crc kubenswrapper[4880]: I0126 16:11:05.305897 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerStarted","Data":"4c8989c2304d3488c07c9cba84a112ba6b5622714014ca2ac091527d1e158143"} Jan 26 16:11:06 crc kubenswrapper[4880]: I0126 16:11:06.332754 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerStarted","Data":"689a187d92622ed751fa815361a73b8392d45c79608bcb00dd9f452dd691226a"} Jan 26 16:11:06 crc kubenswrapper[4880]: I0126 16:11:06.335885 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerStarted","Data":"bc294639b310212e0d653dfc658b8631ab131ca10b08761a51c51413f66f516e"} Jan 26 16:11:12 crc kubenswrapper[4880]: E0126 16:11:12.733149 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" podUID="5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16" Jan 26 16:11:12 crc kubenswrapper[4880]: E0126 16:11:12.733593 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" podUID="04efd524-476a-4803-8996-9b16b3214c32" Jan 26 16:11:12 crc kubenswrapper[4880]: E0126 16:11:12.745970 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" podUID="3e64bef9-c93b-49e3-93ee-4161a2a7c89c" Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.408467 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"d0610ce0803371f0fec80c0a9e973f47bcb6d3a7a472278695c6606d95676ffc"} Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.410598 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"535aedb6aec14a5b49f75204824b2a30ec2d9f6a0b3d38402331408145452414"} Jan 26 16:11:13 crc kubenswrapper[4880]: E0126 16:11:13.411652 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/sg-bridge:latest\\\"\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" podUID="04efd524-476a-4803-8996-9b16b3214c32" Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.412578 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"038f7b3227d1f9d8ea34f0b293f7f9a070419d6ee4c31c598561bca0acbd24f0"} Jan 26 16:11:13 crc kubenswrapper[4880]: E0126 16:11:13.413055 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/sg-bridge:latest\\\"\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" podUID="5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16" Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.415874 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerStarted","Data":"282f2a0e0c16d4c01db38779736eb3875edd0d8429f2da855c8b090665f494a4"} Jan 26 16:11:13 crc kubenswrapper[4880]: E0126 16:11:13.430351 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/sg-bridge:latest\\\"\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" podUID="3e64bef9-c93b-49e3-93ee-4161a2a7c89c" Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.433584 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerStarted","Data":"2d3ab8ed6079363b739d2d0a2d9245d6c98a98128e398421d6979d626e6f96c9"} Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.469161 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" podStartSLOduration=10.126439817 podStartE2EDuration="18.469109872s" podCreationTimestamp="2026-01-26 16:10:55 +0000 UTC" firstStartedPulling="2026-01-26 16:11:04.140591783 +0000 UTC m=+989.626320480" lastFinishedPulling="2026-01-26 16:11:12.483261818 +0000 UTC m=+997.968990535" observedRunningTime="2026-01-26 16:11:13.460801116 +0000 UTC m=+998.946529823" watchObservedRunningTime="2026-01-26 16:11:13.469109872 +0000 UTC m=+998.954838579" Jan 26 16:11:13 crc kubenswrapper[4880]: I0126 16:11:13.499722 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" podStartSLOduration=9.174252075 podStartE2EDuration="17.499703152s" podCreationTimestamp="2026-01-26 16:10:56 +0000 UTC" firstStartedPulling="2026-01-26 16:11:04.17024192 +0000 UTC m=+989.655970627" lastFinishedPulling="2026-01-26 16:11:12.495692977 +0000 UTC m=+997.981421704" observedRunningTime="2026-01-26 16:11:13.498201225 +0000 UTC m=+998.983929932" watchObservedRunningTime="2026-01-26 16:11:13.499703152 +0000 UTC m=+998.985431859" Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.428357 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.428894 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" podUID="6bfefc28-b018-4cd5-b468-cd6f4fede198" containerName="default-interconnect" containerID="cri-o://f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a" gracePeriod=30 Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.487998 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"7c1846568edafda5c151e194357631fc5de249ff4eceb4e31008458ff56ce52f"} Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.494284 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"e1fb4d964c30a9169adde5611d00a70e3d77251e1ec8114857efedad9fea8d09"} Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.501006 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"8d509186595064e1482d15f3118a93989bfa0aaa309b1bf1d10eb01a00a4c669"} Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.579472 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" podStartSLOduration=3.13133502 podStartE2EDuration="29.579448515s" podCreationTimestamp="2026-01-26 16:10:46 +0000 UTC" firstStartedPulling="2026-01-26 16:10:48.405962968 +0000 UTC m=+973.891691675" lastFinishedPulling="2026-01-26 16:11:14.854076463 +0000 UTC m=+1000.339805170" observedRunningTime="2026-01-26 16:11:15.575679911 +0000 UTC m=+1001.061408618" watchObservedRunningTime="2026-01-26 16:11:15.579448515 +0000 UTC m=+1001.065177212" Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.580606 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" podStartSLOduration=4.108820337 podStartE2EDuration="34.580597434s" podCreationTimestamp="2026-01-26 16:10:41 +0000 UTC" firstStartedPulling="2026-01-26 16:10:44.440516033 +0000 UTC m=+969.926244740" lastFinishedPulling="2026-01-26 16:11:14.91229313 +0000 UTC m=+1000.398021837" observedRunningTime="2026-01-26 16:11:15.518036509 +0000 UTC m=+1001.003765216" watchObservedRunningTime="2026-01-26 16:11:15.580597434 +0000 UTC m=+1001.066326141" Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.616451 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" podStartSLOduration=6.103097666 podStartE2EDuration="36.616353232s" podCreationTimestamp="2026-01-26 16:10:39 +0000 UTC" firstStartedPulling="2026-01-26 16:10:44.342819736 +0000 UTC m=+969.828548443" lastFinishedPulling="2026-01-26 16:11:14.856075302 +0000 UTC m=+1000.341804009" observedRunningTime="2026-01-26 16:11:15.609140043 +0000 UTC m=+1001.094868740" watchObservedRunningTime="2026-01-26 16:11:15.616353232 +0000 UTC m=+1001.102081939" Jan 26 16:11:15 crc kubenswrapper[4880]: I0126 16:11:15.857133 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.036976 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037086 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037154 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037215 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq8fm\" (UniqueName: \"kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037244 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037272 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.037299 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config\") pod \"6bfefc28-b018-4cd5-b468-cd6f4fede198\" (UID: \"6bfefc28-b018-4cd5-b468-cd6f4fede198\") " Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.038080 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.043137 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.043261 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.043379 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.043513 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.044325 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.047136 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm" (OuterVolumeSpecName: "kube-api-access-rq8fm") pod "6bfefc28-b018-4cd5-b468-cd6f4fede198" (UID: "6bfefc28-b018-4cd5-b468-cd6f4fede198"). InnerVolumeSpecName "kube-api-access-rq8fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.138934 4880 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.138986 4880 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.138997 4880 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-users\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.139011 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq8fm\" (UniqueName: \"kubernetes.io/projected/6bfefc28-b018-4cd5-b468-cd6f4fede198-kube-api-access-rq8fm\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.139022 4880 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.139035 4880 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/6bfefc28-b018-4cd5-b468-cd6f4fede198-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.139045 4880 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/6bfefc28-b018-4cd5-b468-cd6f4fede198-sasl-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.509224 4880 generic.go:334] "Generic (PLEG): container finished" podID="04efd524-476a-4803-8996-9b16b3214c32" containerID="7c1846568edafda5c151e194357631fc5de249ff4eceb4e31008458ff56ce52f" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.509281 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerDied","Data":"7c1846568edafda5c151e194357631fc5de249ff4eceb4e31008458ff56ce52f"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.510268 4880 scope.go:117] "RemoveContainer" containerID="7c1846568edafda5c151e194357631fc5de249ff4eceb4e31008458ff56ce52f" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.511802 4880 generic.go:334] "Generic (PLEG): container finished" podID="5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16" containerID="e1fb4d964c30a9169adde5611d00a70e3d77251e1ec8114857efedad9fea8d09" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.511860 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerDied","Data":"e1fb4d964c30a9169adde5611d00a70e3d77251e1ec8114857efedad9fea8d09"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.513922 4880 scope.go:117] "RemoveContainer" containerID="e1fb4d964c30a9169adde5611d00a70e3d77251e1ec8114857efedad9fea8d09" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.517999 4880 generic.go:334] "Generic (PLEG): container finished" podID="3e64bef9-c93b-49e3-93ee-4161a2a7c89c" containerID="8d509186595064e1482d15f3118a93989bfa0aaa309b1bf1d10eb01a00a4c669" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.518044 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerDied","Data":"8d509186595064e1482d15f3118a93989bfa0aaa309b1bf1d10eb01a00a4c669"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.519865 4880 generic.go:334] "Generic (PLEG): container finished" podID="71cb321c-ed35-4d06-a487-eacea2dee247" containerID="bc294639b310212e0d653dfc658b8631ab131ca10b08761a51c51413f66f516e" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.519934 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerDied","Data":"bc294639b310212e0d653dfc658b8631ab131ca10b08761a51c51413f66f516e"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.520500 4880 scope.go:117] "RemoveContainer" containerID="bc294639b310212e0d653dfc658b8631ab131ca10b08761a51c51413f66f516e" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.520739 4880 scope.go:117] "RemoveContainer" containerID="8d509186595064e1482d15f3118a93989bfa0aaa309b1bf1d10eb01a00a4c669" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.526305 4880 generic.go:334] "Generic (PLEG): container finished" podID="e07f674d-87fb-46ce-98bc-34c04cf788da" containerID="689a187d92622ed751fa815361a73b8392d45c79608bcb00dd9f452dd691226a" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.526378 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerDied","Data":"689a187d92622ed751fa815361a73b8392d45c79608bcb00dd9f452dd691226a"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.527018 4880 scope.go:117] "RemoveContainer" containerID="689a187d92622ed751fa815361a73b8392d45c79608bcb00dd9f452dd691226a" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.532992 4880 generic.go:334] "Generic (PLEG): container finished" podID="6bfefc28-b018-4cd5-b468-cd6f4fede198" containerID="f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a" exitCode=0 Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.533046 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" event={"ID":"6bfefc28-b018-4cd5-b468-cd6f4fede198","Type":"ContainerDied","Data":"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.533081 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" event={"ID":"6bfefc28-b018-4cd5-b468-cd6f4fede198","Type":"ContainerDied","Data":"471c0a57da0df6462858939ff7e767ffb51b3915f8893a27364afd7b10794d1e"} Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.533101 4880 scope.go:117] "RemoveContainer" containerID="f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.533212 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-v69cb" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.566287 4880 scope.go:117] "RemoveContainer" containerID="f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a" Jan 26 16:11:16 crc kubenswrapper[4880]: E0126 16:11:16.568981 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a\": container with ID starting with f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a not found: ID does not exist" containerID="f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.569024 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a"} err="failed to get container status \"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a\": rpc error: code = NotFound desc = could not find container \"f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a\": container with ID starting with f5bcc79612e015e127f5d846d5bcb615a0111f8a64a9a801d1d91c592d1ee21a not found: ID does not exist" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.688529 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.690749 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-v69cb"] Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.823940 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4vlzh"] Jan 26 16:11:16 crc kubenswrapper[4880]: E0126 16:11:16.824235 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bfefc28-b018-4cd5-b468-cd6f4fede198" containerName="default-interconnect" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.824247 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bfefc28-b018-4cd5-b468-cd6f4fede198" containerName="default-interconnect" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.824391 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bfefc28-b018-4cd5-b468-cd6f4fede198" containerName="default-interconnect" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.824904 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.827247 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.827394 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.827453 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-pbkdw" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.827523 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.827792 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.829098 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.829700 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.845018 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4vlzh"] Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951594 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-config\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951674 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951708 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951753 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6d9s\" (UniqueName: \"kubernetes.io/projected/728e530c-af3d-417b-912d-ce69f8b6b0b8-kube-api-access-p6d9s\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951784 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-users\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951825 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:16 crc kubenswrapper[4880]: I0126 16:11:16.951853 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053214 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-config\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053300 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053331 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053378 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6d9s\" (UniqueName: \"kubernetes.io/projected/728e530c-af3d-417b-912d-ce69f8b6b0b8-kube-api-access-p6d9s\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053412 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-users\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053468 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.053492 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.055205 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-config\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.069841 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-sasl-users\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.070174 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.072121 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.072659 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.078705 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6d9s\" (UniqueName: \"kubernetes.io/projected/728e530c-af3d-417b-912d-ce69f8b6b0b8-kube-api-access-p6d9s\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.080648 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/728e530c-af3d-417b-912d-ce69f8b6b0b8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-4vlzh\" (UID: \"728e530c-af3d-417b-912d-ce69f8b6b0b8\") " pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.152684 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.541549 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"a88e6fad8f7a2c0b37806c22a0fa3b83754484a5acb43f0c515c51eb05be8fa9"} Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.554746 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"70ce6d39c3038315701197bec8fe317626ecf7785f38cf81501190fe1c04536b"} Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.566703 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerStarted","Data":"a4739c02da01838707480ebe63a3b8ccd9eb2e8b794c53b474f2946dedf619f4"} Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.571569 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerStarted","Data":"992674321d42531b1ad7e45d4f01bae027704cc9550031b04b871956297c2bd7"} Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.578910 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"62c64da079ee8da66d9a408104481e85627198d971af5ee43dd5551bb5d286b6"} Jan 26 16:11:17 crc kubenswrapper[4880]: I0126 16:11:17.701305 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-4vlzh"] Jan 26 16:11:17 crc kubenswrapper[4880]: W0126 16:11:17.711058 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod728e530c_af3d_417b_912d_ce69f8b6b0b8.slice/crio-7d48bb29ca274858b14b7bd3c667630a92336c4f7e56d4e7439027e2c922bcfc WatchSource:0}: Error finding container 7d48bb29ca274858b14b7bd3c667630a92336c4f7e56d4e7439027e2c922bcfc: Status 404 returned error can't find the container with id 7d48bb29ca274858b14b7bd3c667630a92336c4f7e56d4e7439027e2c922bcfc Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.223405 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bfefc28-b018-4cd5-b468-cd6f4fede198" path="/var/lib/kubelet/pods/6bfefc28-b018-4cd5-b468-cd6f4fede198/volumes" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.591168 4880 generic.go:334] "Generic (PLEG): container finished" podID="e07f674d-87fb-46ce-98bc-34c04cf788da" containerID="992674321d42531b1ad7e45d4f01bae027704cc9550031b04b871956297c2bd7" exitCode=0 Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.591235 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerDied","Data":"992674321d42531b1ad7e45d4f01bae027704cc9550031b04b871956297c2bd7"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.591271 4880 scope.go:117] "RemoveContainer" containerID="689a187d92622ed751fa815361a73b8392d45c79608bcb00dd9f452dd691226a" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.592016 4880 scope.go:117] "RemoveContainer" containerID="992674321d42531b1ad7e45d4f01bae027704cc9550031b04b871956297c2bd7" Jan 26 16:11:18 crc kubenswrapper[4880]: E0126 16:11:18.592488 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd_service-telemetry(e07f674d-87fb-46ce-98bc-34c04cf788da)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" podUID="e07f674d-87fb-46ce-98bc-34c04cf788da" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.610095 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" event={"ID":"728e530c-af3d-417b-912d-ce69f8b6b0b8","Type":"ContainerStarted","Data":"af71067529d76523fd57d8073eac88ffc99616752120d25687aa9fffaf3500fc"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.610152 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" event={"ID":"728e530c-af3d-417b-912d-ce69f8b6b0b8","Type":"ContainerStarted","Data":"7d48bb29ca274858b14b7bd3c667630a92336c4f7e56d4e7439027e2c922bcfc"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.621608 4880 generic.go:334] "Generic (PLEG): container finished" podID="04efd524-476a-4803-8996-9b16b3214c32" containerID="62c64da079ee8da66d9a408104481e85627198d971af5ee43dd5551bb5d286b6" exitCode=0 Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.621699 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerDied","Data":"62c64da079ee8da66d9a408104481e85627198d971af5ee43dd5551bb5d286b6"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.622593 4880 scope.go:117] "RemoveContainer" containerID="62c64da079ee8da66d9a408104481e85627198d971af5ee43dd5551bb5d286b6" Jan 26 16:11:18 crc kubenswrapper[4880]: E0126 16:11:18.622843 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh_service-telemetry(04efd524-476a-4803-8996-9b16b3214c32)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" podUID="04efd524-476a-4803-8996-9b16b3214c32" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.635982 4880 generic.go:334] "Generic (PLEG): container finished" podID="5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16" containerID="a88e6fad8f7a2c0b37806c22a0fa3b83754484a5acb43f0c515c51eb05be8fa9" exitCode=0 Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.636072 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerDied","Data":"a88e6fad8f7a2c0b37806c22a0fa3b83754484a5acb43f0c515c51eb05be8fa9"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.636819 4880 scope.go:117] "RemoveContainer" containerID="a88e6fad8f7a2c0b37806c22a0fa3b83754484a5acb43f0c515c51eb05be8fa9" Jan 26 16:11:18 crc kubenswrapper[4880]: E0126 16:11:18.637058 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4_service-telemetry(5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" podUID="5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.652653 4880 scope.go:117] "RemoveContainer" containerID="7c1846568edafda5c151e194357631fc5de249ff4eceb4e31008458ff56ce52f" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.663181 4880 generic.go:334] "Generic (PLEG): container finished" podID="3e64bef9-c93b-49e3-93ee-4161a2a7c89c" containerID="70ce6d39c3038315701197bec8fe317626ecf7785f38cf81501190fe1c04536b" exitCode=0 Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.663292 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerDied","Data":"70ce6d39c3038315701197bec8fe317626ecf7785f38cf81501190fe1c04536b"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.664091 4880 scope.go:117] "RemoveContainer" containerID="70ce6d39c3038315701197bec8fe317626ecf7785f38cf81501190fe1c04536b" Jan 26 16:11:18 crc kubenswrapper[4880]: E0126 16:11:18.664326 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt_service-telemetry(3e64bef9-c93b-49e3-93ee-4161a2a7c89c)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" podUID="3e64bef9-c93b-49e3-93ee-4161a2a7c89c" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.677326 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-4vlzh" podStartSLOduration=3.6773068159999998 podStartE2EDuration="3.677306816s" podCreationTimestamp="2026-01-26 16:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:11:18.671208983 +0000 UTC m=+1004.156937700" watchObservedRunningTime="2026-01-26 16:11:18.677306816 +0000 UTC m=+1004.163035523" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.689221 4880 generic.go:334] "Generic (PLEG): container finished" podID="71cb321c-ed35-4d06-a487-eacea2dee247" containerID="a4739c02da01838707480ebe63a3b8ccd9eb2e8b794c53b474f2946dedf619f4" exitCode=0 Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.689276 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerDied","Data":"a4739c02da01838707480ebe63a3b8ccd9eb2e8b794c53b474f2946dedf619f4"} Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.689922 4880 scope.go:117] "RemoveContainer" containerID="a4739c02da01838707480ebe63a3b8ccd9eb2e8b794c53b474f2946dedf619f4" Jan 26 16:11:18 crc kubenswrapper[4880]: E0126 16:11:18.690194 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-76756967c5-bpljr_service-telemetry(71cb321c-ed35-4d06-a487-eacea2dee247)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" podUID="71cb321c-ed35-4d06-a487-eacea2dee247" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.728285 4880 scope.go:117] "RemoveContainer" containerID="e1fb4d964c30a9169adde5611d00a70e3d77251e1ec8114857efedad9fea8d09" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.777907 4880 scope.go:117] "RemoveContainer" containerID="8d509186595064e1482d15f3118a93989bfa0aaa309b1bf1d10eb01a00a4c669" Jan 26 16:11:18 crc kubenswrapper[4880]: I0126 16:11:18.822472 4880 scope.go:117] "RemoveContainer" containerID="bc294639b310212e0d653dfc658b8631ab131ca10b08761a51c51413f66f516e" Jan 26 16:11:29 crc kubenswrapper[4880]: I0126 16:11:29.213718 4880 scope.go:117] "RemoveContainer" containerID="a4739c02da01838707480ebe63a3b8ccd9eb2e8b794c53b474f2946dedf619f4" Jan 26 16:11:29 crc kubenswrapper[4880]: I0126 16:11:29.215168 4880 scope.go:117] "RemoveContainer" containerID="70ce6d39c3038315701197bec8fe317626ecf7785f38cf81501190fe1c04536b" Jan 26 16:11:29 crc kubenswrapper[4880]: I0126 16:11:29.215463 4880 scope.go:117] "RemoveContainer" containerID="a88e6fad8f7a2c0b37806c22a0fa3b83754484a5acb43f0c515c51eb05be8fa9" Jan 26 16:11:30 crc kubenswrapper[4880]: I0126 16:11:30.795530 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4" event={"ID":"5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16","Type":"ContainerStarted","Data":"62a93666ac1616aa096a48940bebe0852b1307f16bc0c9d4417180002bee08a6"} Jan 26 16:11:30 crc kubenswrapper[4880]: I0126 16:11:30.797925 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt" event={"ID":"3e64bef9-c93b-49e3-93ee-4161a2a7c89c","Type":"ContainerStarted","Data":"2eef85242208a7644a44a72c85f4f8c3b2447792d69e32d302144902bdec1734"} Jan 26 16:11:30 crc kubenswrapper[4880]: I0126 16:11:30.800647 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-76756967c5-bpljr" event={"ID":"71cb321c-ed35-4d06-a487-eacea2dee247","Type":"ContainerStarted","Data":"c4c90c4fe413ca4d9ad3af0ad6761ad1c0d3e789111603ddfbd4cc3c3db57d44"} Jan 26 16:11:32 crc kubenswrapper[4880]: I0126 16:11:32.214578 4880 scope.go:117] "RemoveContainer" containerID="62c64da079ee8da66d9a408104481e85627198d971af5ee43dd5551bb5d286b6" Jan 26 16:11:33 crc kubenswrapper[4880]: I0126 16:11:33.215401 4880 scope.go:117] "RemoveContainer" containerID="992674321d42531b1ad7e45d4f01bae027704cc9550031b04b871956297c2bd7" Jan 26 16:11:33 crc kubenswrapper[4880]: I0126 16:11:33.823178 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh" event={"ID":"04efd524-476a-4803-8996-9b16b3214c32","Type":"ContainerStarted","Data":"9d03b46f9d7421abda018282fce03d80a4ec5e17f7ed0f2fd6ce1b1706cc64f4"} Jan 26 16:11:34 crc kubenswrapper[4880]: I0126 16:11:34.831474 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd" event={"ID":"e07f674d-87fb-46ce-98bc-34c04cf788da","Type":"ContainerStarted","Data":"9b3959cf6b154a5c1ac80f7240c145cf1529d5275303fc71161238a79418058d"} Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.523305 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.525325 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.528524 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.529131 4880 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.571674 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.747023 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw98h\" (UniqueName: \"kubernetes.io/projected/23a071f2-a160-4871-bc14-cae91d0c08c9-kube-api-access-mw98h\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.747086 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/23a071f2-a160-4871-bc14-cae91d0c08c9-qdr-test-config\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.747171 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/23a071f2-a160-4871-bc14-cae91d0c08c9-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.849367 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/23a071f2-a160-4871-bc14-cae91d0c08c9-qdr-test-config\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.849500 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/23a071f2-a160-4871-bc14-cae91d0c08c9-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.849606 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw98h\" (UniqueName: \"kubernetes.io/projected/23a071f2-a160-4871-bc14-cae91d0c08c9-kube-api-access-mw98h\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.850414 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/23a071f2-a160-4871-bc14-cae91d0c08c9-qdr-test-config\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.858394 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/23a071f2-a160-4871-bc14-cae91d0c08c9-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:45 crc kubenswrapper[4880]: I0126 16:11:45.870983 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw98h\" (UniqueName: \"kubernetes.io/projected/23a071f2-a160-4871-bc14-cae91d0c08c9-kube-api-access-mw98h\") pod \"qdr-test\" (UID: \"23a071f2-a160-4871-bc14-cae91d0c08c9\") " pod="service-telemetry/qdr-test" Jan 26 16:11:46 crc kubenswrapper[4880]: I0126 16:11:46.167765 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Jan 26 16:11:46 crc kubenswrapper[4880]: I0126 16:11:46.569394 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Jan 26 16:11:46 crc kubenswrapper[4880]: W0126 16:11:46.575081 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a071f2_a160_4871_bc14_cae91d0c08c9.slice/crio-667afa854cd681d97273b566ff04ff3a11e88596b586beba61f309a37c0ae336 WatchSource:0}: Error finding container 667afa854cd681d97273b566ff04ff3a11e88596b586beba61f309a37c0ae336: Status 404 returned error can't find the container with id 667afa854cd681d97273b566ff04ff3a11e88596b586beba61f309a37c0ae336 Jan 26 16:11:47 crc kubenswrapper[4880]: I0126 16:11:47.441931 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"23a071f2-a160-4871-bc14-cae91d0c08c9","Type":"ContainerStarted","Data":"667afa854cd681d97273b566ff04ff3a11e88596b586beba61f309a37c0ae336"} Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.521690 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"23a071f2-a160-4871-bc14-cae91d0c08c9","Type":"ContainerStarted","Data":"b69e3accb82b0e15bed009c1439fc8c2e3c3090e6437eac2082d0e5a03a019cf"} Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.558020 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.716717366 podStartE2EDuration="11.55795375s" podCreationTimestamp="2026-01-26 16:11:45 +0000 UTC" firstStartedPulling="2026-01-26 16:11:46.577626519 +0000 UTC m=+1032.063355226" lastFinishedPulling="2026-01-26 16:11:55.418862903 +0000 UTC m=+1040.904591610" observedRunningTime="2026-01-26 16:11:56.542804796 +0000 UTC m=+1042.028533503" watchObservedRunningTime="2026-01-26 16:11:56.55795375 +0000 UTC m=+1042.043682497" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.812902 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-vvzvt"] Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.814172 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.816708 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.817255 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.817463 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.817633 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.817677 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.818718 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.828312 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-vvzvt"] Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.879790 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.879871 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.879959 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981466 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981771 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981818 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981878 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981904 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981934 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.981968 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24w7f\" (UniqueName: \"kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.982757 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.982952 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:56 crc kubenswrapper[4880]: I0126 16:11:56.983101 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.082813 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.083125 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.083228 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.083343 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24w7f\" (UniqueName: \"kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.083994 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.084089 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.084321 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.105101 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24w7f\" (UniqueName: \"kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f\") pod \"stf-smoketest-smoke1-vvzvt\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.132659 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.166820 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.167710 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.185403 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.196856 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqfs2\" (UniqueName: \"kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2\") pod \"curl\" (UID: \"c53126d4-6b4a-41d7-a4d6-5deb65a0e876\") " pod="service-telemetry/curl" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.300378 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqfs2\" (UniqueName: \"kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2\") pod \"curl\" (UID: \"c53126d4-6b4a-41d7-a4d6-5deb65a0e876\") " pod="service-telemetry/curl" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.344301 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqfs2\" (UniqueName: \"kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2\") pod \"curl\" (UID: \"c53126d4-6b4a-41d7-a4d6-5deb65a0e876\") " pod="service-telemetry/curl" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.537709 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.649495 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-vvzvt"] Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.666968 4880 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 16:11:57 crc kubenswrapper[4880]: I0126 16:11:57.789368 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Jan 26 16:11:57 crc kubenswrapper[4880]: W0126 16:11:57.789674 4880 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc53126d4_6b4a_41d7_a4d6_5deb65a0e876.slice/crio-8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451 WatchSource:0}: Error finding container 8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451: Status 404 returned error can't find the container with id 8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451 Jan 26 16:11:58 crc kubenswrapper[4880]: I0126 16:11:58.536976 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerStarted","Data":"013b25dfb64f72487883d89dd229bbf465f751303edd38faafb5d010c1e6006f"} Jan 26 16:11:58 crc kubenswrapper[4880]: I0126 16:11:58.538052 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"c53126d4-6b4a-41d7-a4d6-5deb65a0e876","Type":"ContainerStarted","Data":"8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451"} Jan 26 16:11:59 crc kubenswrapper[4880]: I0126 16:11:59.547996 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"c53126d4-6b4a-41d7-a4d6-5deb65a0e876","Type":"ContainerStarted","Data":"c334ca920cf9fc26ed7e2e21b212e5b975f2a3423ae0201040e784c4cf9fd894"} Jan 26 16:11:59 crc kubenswrapper[4880]: I0126 16:11:59.571288 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/curl" podStartSLOduration=0.995588232 podStartE2EDuration="2.571266788s" podCreationTimestamp="2026-01-26 16:11:57 +0000 UTC" firstStartedPulling="2026-01-26 16:11:57.79230519 +0000 UTC m=+1043.278033907" lastFinishedPulling="2026-01-26 16:11:59.367983766 +0000 UTC m=+1044.853712463" observedRunningTime="2026-01-26 16:11:59.564823709 +0000 UTC m=+1045.050552416" watchObservedRunningTime="2026-01-26 16:11:59.571266788 +0000 UTC m=+1045.056995495" Jan 26 16:12:00 crc kubenswrapper[4880]: I0126 16:12:00.555958 4880 generic.go:334] "Generic (PLEG): container finished" podID="c53126d4-6b4a-41d7-a4d6-5deb65a0e876" containerID="c334ca920cf9fc26ed7e2e21b212e5b975f2a3423ae0201040e784c4cf9fd894" exitCode=0 Jan 26 16:12:00 crc kubenswrapper[4880]: I0126 16:12:00.555999 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"c53126d4-6b4a-41d7-a4d6-5deb65a0e876","Type":"ContainerDied","Data":"c334ca920cf9fc26ed7e2e21b212e5b975f2a3423ae0201040e784c4cf9fd894"} Jan 26 16:12:09 crc kubenswrapper[4880]: I0126 16:12:09.812848 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 26 16:12:09 crc kubenswrapper[4880]: I0126 16:12:09.910750 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqfs2\" (UniqueName: \"kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2\") pod \"c53126d4-6b4a-41d7-a4d6-5deb65a0e876\" (UID: \"c53126d4-6b4a-41d7-a4d6-5deb65a0e876\") " Jan 26 16:12:09 crc kubenswrapper[4880]: I0126 16:12:09.914284 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2" (OuterVolumeSpecName: "kube-api-access-cqfs2") pod "c53126d4-6b4a-41d7-a4d6-5deb65a0e876" (UID: "c53126d4-6b4a-41d7-a4d6-5deb65a0e876"). InnerVolumeSpecName "kube-api-access-cqfs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:12:09 crc kubenswrapper[4880]: I0126 16:12:09.974672 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_c53126d4-6b4a-41d7-a4d6-5deb65a0e876/curl/0.log" Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.012616 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqfs2\" (UniqueName: \"kubernetes.io/projected/c53126d4-6b4a-41d7-a4d6-5deb65a0e876-kube-api-access-cqfs2\") on node \"crc\" DevicePath \"\"" Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.232382 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-j5sr8_edc9710b-d4f1-4a5a-b434-2e471dc1a491/prometheus-webhook-snmp/0.log" Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.631152 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerStarted","Data":"4f7fbe466631c393d64ee1940e5a77a4e142888dac3d8dd387f0852a7d34182b"} Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.633182 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"c53126d4-6b4a-41d7-a4d6-5deb65a0e876","Type":"ContainerDied","Data":"8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451"} Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.633231 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cc14e0253c71f4717e1cd116c5447db2296a8702e5300e8857d1204a0e1f451" Jan 26 16:12:10 crc kubenswrapper[4880]: I0126 16:12:10.633334 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Jan 26 16:12:23 crc kubenswrapper[4880]: E0126 16:12:23.371005 4880 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo" Jan 26 16:12:23 crc kubenswrapper[4880]: E0126 16:12:23.371911 4880 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:smoketest-ceilometer,Image:quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo,Command:[/smoketest_ceilometer_entrypoint.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CLOUDNAME,Value:smoke1,ValueFrom:nil,},EnvVar{Name:ELASTICSEARCH_AUTH_PASS,Value:vdGWXNNq1SIkkOaYu9cvwaMq,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_AUTH_TOKEN,Value:eyJhbGciOiJSUzI1NiIsImtpZCI6InF6SnFxNFFjbVk5VmJQZ2dNMmUxdHFmTlJlVWx4UDhSTlhIamV3RUx4WU0ifQ.eyJhdWQiOlsiaHR0cHM6Ly9rdWJlcm5ldGVzLmRlZmF1bHQuc3ZjIl0sImV4cCI6MTc2OTQ0NzUwMCwiaWF0IjoxNzY5NDQzOTAwLCJpc3MiOiJodHRwczovL2t1YmVybmV0ZXMuZGVmYXVsdC5zdmMiLCJqdGkiOiI5YzE0YzBhOS03MzdiLTQ0MWEtODMwYy0yNmRiZjEyYzExNTUiLCJrdWJlcm5ldGVzLmlvIjp7Im5hbWVzcGFjZSI6InNlcnZpY2UtdGVsZW1ldHJ5Iiwic2VydmljZWFjY291bnQiOnsibmFtZSI6InN0Zi1wcm9tZXRoZXVzLXJlYWRlciIsInVpZCI6ImM1YWI0NmEzLTdiZGUtNDU0Zi1hZTMwLWZhYWQyNzgwMzFlYiJ9fSwibmJmIjoxNzY5NDQzOTAwLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6c2VydmljZS10ZWxlbWV0cnk6c3RmLXByb21ldGhldXMtcmVhZGVyIn0.pBuzyGP1l2T9rj2jIXqwN-KEOmjFvZhNiaSM8DFWSgILXXko5coeBWHQHzBGdqGlfkq5EDmMXp-To73W1pXp_Lzu8jWxy7TO63RG5Q5qpJKAPz7bFzQjLL_yHPkIqSu6dMLkwS5qzQcNzdrvYpAOTzMcFjh3tACcClxC4IlEXKbwIqIMjzADuJ01AJ6sel1oEcpvfJKoZGHVC4wCCqqd_pFDUP22DLBXrf3MbWkq5P35omGbDifcuKfLSbv0PYXuZxeqaFFH8WL7Inhbkd3BaoGGZZsW-eNM4lg1lZ8oI20PcNRFQ1bURhGIgAn-YVqSTcorzXezcshesMA9dqwOy_Ih_ag1SeLUhjNke9nESCxaWd1e1Pf68vHbpFDgTkqXvppJWkWn7gSC1V25iVBissV-drZTLp1LU2UmObkWdGo7bt7LlVfXqZGgOuxr_vf9ow73BDqKlZhE9R9bqwExODDxGoXz2DPzm9BO0PfDweYljWioLLMaEdKIlyEqjszra-NVBbKdahm2cKOHua3NXYAivA9nmwWyryBZxyXOFD3ZJR_JwfDZYe6P4AvQRJML8WleH5dDZxSGEQ4AQA-a-5yu5CTC147WoqxWMmx3NqZG06WUbQ6_WUdABbn62z3Rl7_TYBYg0Xjk0PGtqjCcw6x0cRRs9qlZwNTQORn61OQ,ValueFrom:nil,},EnvVar{Name:OBSERVABILITY_STRATEGY,Value:<>,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ceilometer-publisher,ReadOnly:false,MountPath:/ceilometer_publish.py,SubPath:ceilometer_publish.py,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceilometer-entrypoint-script,ReadOnly:false,MountPath:/smoketest_ceilometer_entrypoint.sh,SubPath:smoketest_ceilometer_entrypoint.sh,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24w7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod stf-smoketest-smoke1-vvzvt_service-telemetry(5c8ea08b-1b0a-4f44-9915-ef222a39783f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 16:12:23 crc kubenswrapper[4880]: E0126 16:12:23.373218 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"smoketest-ceilometer\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" Jan 26 16:12:23 crc kubenswrapper[4880]: E0126 16:12:23.881040 4880 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"smoketest-ceilometer\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/tripleomastercentos9/openstack-ceilometer-notification:current-tripleo\\\"\"" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" Jan 26 16:12:34 crc kubenswrapper[4880]: I0126 16:12:34.992029 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerStarted","Data":"6c279f933838442c858c1954516f9d522af94bd7a7d1e1c4e992d8141ecb2aeb"} Jan 26 16:12:35 crc kubenswrapper[4880]: I0126 16:12:35.014776 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" podStartSLOduration=2.021163745 podStartE2EDuration="39.014731438s" podCreationTimestamp="2026-01-26 16:11:56 +0000 UTC" firstStartedPulling="2026-01-26 16:11:57.666562363 +0000 UTC m=+1043.152291080" lastFinishedPulling="2026-01-26 16:12:34.660130066 +0000 UTC m=+1080.145858773" observedRunningTime="2026-01-26 16:12:35.010010832 +0000 UTC m=+1080.495739549" watchObservedRunningTime="2026-01-26 16:12:35.014731438 +0000 UTC m=+1080.500460145" Jan 26 16:12:40 crc kubenswrapper[4880]: I0126 16:12:40.369977 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-j5sr8_edc9710b-d4f1-4a5a-b434-2e471dc1a491/prometheus-webhook-snmp/0.log" Jan 26 16:12:45 crc kubenswrapper[4880]: I0126 16:12:45.075892 4880 generic.go:334] "Generic (PLEG): container finished" podID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerID="4f7fbe466631c393d64ee1940e5a77a4e142888dac3d8dd387f0852a7d34182b" exitCode=0 Jan 26 16:12:45 crc kubenswrapper[4880]: I0126 16:12:45.075969 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerDied","Data":"4f7fbe466631c393d64ee1940e5a77a4e142888dac3d8dd387f0852a7d34182b"} Jan 26 16:12:45 crc kubenswrapper[4880]: I0126 16:12:45.077106 4880 scope.go:117] "RemoveContainer" containerID="4f7fbe466631c393d64ee1940e5a77a4e142888dac3d8dd387f0852a7d34182b" Jan 26 16:13:07 crc kubenswrapper[4880]: I0126 16:13:07.268730 4880 generic.go:334] "Generic (PLEG): container finished" podID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerID="6c279f933838442c858c1954516f9d522af94bd7a7d1e1c4e992d8141ecb2aeb" exitCode=0 Jan 26 16:13:07 crc kubenswrapper[4880]: I0126 16:13:07.268823 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerDied","Data":"6c279f933838442c858c1954516f9d522af94bd7a7d1e1c4e992d8141ecb2aeb"} Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.546512 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.721979 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722067 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722101 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722153 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722200 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722297 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.722358 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24w7f\" (UniqueName: \"kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f\") pod \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\" (UID: \"5c8ea08b-1b0a-4f44-9915-ef222a39783f\") " Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.728771 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f" (OuterVolumeSpecName: "kube-api-access-24w7f") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "kube-api-access-24w7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.741978 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.744812 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.747057 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.747381 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.755046 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.756054 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "5c8ea08b-1b0a-4f44-9915-ef222a39783f" (UID: "5c8ea08b-1b0a-4f44-9915-ef222a39783f"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.824933 4880 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-healthcheck-log\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.824999 4880 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.825016 4880 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.825029 4880 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.825041 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24w7f\" (UniqueName: \"kubernetes.io/projected/5c8ea08b-1b0a-4f44-9915-ef222a39783f-kube-api-access-24w7f\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.825054 4880 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-sensubility-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:08 crc kubenswrapper[4880]: I0126 16:13:08.825064 4880 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5c8ea08b-1b0a-4f44-9915-ef222a39783f-collectd-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:09 crc kubenswrapper[4880]: I0126 16:13:09.288047 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" event={"ID":"5c8ea08b-1b0a-4f44-9915-ef222a39783f","Type":"ContainerDied","Data":"013b25dfb64f72487883d89dd229bbf465f751303edd38faafb5d010c1e6006f"} Jan 26 16:13:09 crc kubenswrapper[4880]: I0126 16:13:09.288107 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="013b25dfb64f72487883d89dd229bbf465f751303edd38faafb5d010c1e6006f" Jan 26 16:13:09 crc kubenswrapper[4880]: I0126 16:13:09.288187 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-vvzvt" Jan 26 16:13:10 crc kubenswrapper[4880]: I0126 16:13:10.486834 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-vvzvt_5c8ea08b-1b0a-4f44-9915-ef222a39783f/smoketest-collectd/0.log" Jan 26 16:13:10 crc kubenswrapper[4880]: I0126 16:13:10.699828 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:13:10 crc kubenswrapper[4880]: I0126 16:13:10.699902 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:13:10 crc kubenswrapper[4880]: I0126 16:13:10.752011 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-vvzvt_5c8ea08b-1b0a-4f44-9915-ef222a39783f/smoketest-ceilometer/0.log" Jan 26 16:13:11 crc kubenswrapper[4880]: I0126 16:13:11.000719 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-4vlzh_728e530c-af3d-417b-912d-ce69f8b6b0b8/default-interconnect/0.log" Jan 26 16:13:11 crc kubenswrapper[4880]: I0126 16:13:11.246769 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt_3e64bef9-c93b-49e3-93ee-4161a2a7c89c/bridge/2.log" Jan 26 16:13:11 crc kubenswrapper[4880]: I0126 16:13:11.624558 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7996dc9458-w22lt_3e64bef9-c93b-49e3-93ee-4161a2a7c89c/sg-core/0.log" Jan 26 16:13:11 crc kubenswrapper[4880]: I0126 16:13:11.900997 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-76756967c5-bpljr_71cb321c-ed35-4d06-a487-eacea2dee247/bridge/2.log" Jan 26 16:13:12 crc kubenswrapper[4880]: I0126 16:13:12.179071 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-76756967c5-bpljr_71cb321c-ed35-4d06-a487-eacea2dee247/sg-core/0.log" Jan 26 16:13:12 crc kubenswrapper[4880]: I0126 16:13:12.430245 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh_04efd524-476a-4803-8996-9b16b3214c32/bridge/2.log" Jan 26 16:13:12 crc kubenswrapper[4880]: I0126 16:13:12.695740 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-b57f974ff-x9pfh_04efd524-476a-4803-8996-9b16b3214c32/sg-core/0.log" Jan 26 16:13:12 crc kubenswrapper[4880]: I0126 16:13:12.956305 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd_e07f674d-87fb-46ce-98bc-34c04cf788da/bridge/2.log" Jan 26 16:13:13 crc kubenswrapper[4880]: I0126 16:13:13.235052 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-f69bb45c4-94rzd_e07f674d-87fb-46ce-98bc-34c04cf788da/sg-core/0.log" Jan 26 16:13:13 crc kubenswrapper[4880]: I0126 16:13:13.479028 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4_5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16/bridge/2.log" Jan 26 16:13:13 crc kubenswrapper[4880]: I0126 16:13:13.761868 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-6864f4fb65-vngg4_5edcc1e8-51ca-4cc8-9de1-c3966d7bdb16/sg-core/0.log" Jan 26 16:13:17 crc kubenswrapper[4880]: I0126 16:13:17.073574 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-77c9d9f969-8jlvw_695785ce-59d7-4136-83c3-70ad2ab08f8b/operator/0.log" Jan 26 16:13:17 crc kubenswrapper[4880]: I0126 16:13:17.319194 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_77ec4615-159f-4075-84f0-4d2515636f1f/prometheus/0.log" Jan 26 16:13:17 crc kubenswrapper[4880]: I0126 16:13:17.591996 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_fff0a88a-97cf-45a6-8de4-4568e40bac22/elasticsearch/0.log" Jan 26 16:13:17 crc kubenswrapper[4880]: I0126 16:13:17.854671 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-78bcbbdcff-j5sr8_edc9710b-d4f1-4a5a-b434-2e471dc1a491/prometheus-webhook-snmp/0.log" Jan 26 16:13:18 crc kubenswrapper[4880]: I0126 16:13:18.091144 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_a784e60c-653f-4125-8162-ec08686dee59/alertmanager/0.log" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.066958 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:19 crc kubenswrapper[4880]: E0126 16:13:19.068271 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53126d4-6b4a-41d7-a4d6-5deb65a0e876" containerName="curl" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.068409 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53126d4-6b4a-41d7-a4d6-5deb65a0e876" containerName="curl" Jan 26 16:13:19 crc kubenswrapper[4880]: E0126 16:13:19.068524 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-collectd" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.068535 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-collectd" Jan 26 16:13:19 crc kubenswrapper[4880]: E0126 16:13:19.068584 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-ceilometer" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.068594 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-ceilometer" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.069051 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-ceilometer" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.069087 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53126d4-6b4a-41d7-a4d6-5deb65a0e876" containerName="curl" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.069112 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8ea08b-1b0a-4f44-9915-ef222a39783f" containerName="smoketest-collectd" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.069868 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.105653 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.227471 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2rqp\" (UniqueName: \"kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp\") pod \"infrawatch-operators-lrwwt\" (UID: \"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3\") " pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.328510 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2rqp\" (UniqueName: \"kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp\") pod \"infrawatch-operators-lrwwt\" (UID: \"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3\") " pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.350761 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2rqp\" (UniqueName: \"kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp\") pod \"infrawatch-operators-lrwwt\" (UID: \"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3\") " pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.390722 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:19 crc kubenswrapper[4880]: I0126 16:13:19.808560 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:20 crc kubenswrapper[4880]: I0126 16:13:20.411081 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-lrwwt" event={"ID":"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3","Type":"ContainerStarted","Data":"e3477c1f9dcb915a4bad3bd6c8d8caf41e88b4a189c8e790d4acb322e779c248"} Jan 26 16:13:21 crc kubenswrapper[4880]: I0126 16:13:21.420274 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-lrwwt" event={"ID":"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3","Type":"ContainerStarted","Data":"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414"} Jan 26 16:13:21 crc kubenswrapper[4880]: I0126 16:13:21.440183 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-lrwwt" podStartSLOduration=1.9734988759999998 podStartE2EDuration="2.438809355s" podCreationTimestamp="2026-01-26 16:13:19 +0000 UTC" firstStartedPulling="2026-01-26 16:13:19.820063325 +0000 UTC m=+1125.305792032" lastFinishedPulling="2026-01-26 16:13:20.285373804 +0000 UTC m=+1125.771102511" observedRunningTime="2026-01-26 16:13:21.435873313 +0000 UTC m=+1126.921602030" watchObservedRunningTime="2026-01-26 16:13:21.438809355 +0000 UTC m=+1126.924538062" Jan 26 16:13:29 crc kubenswrapper[4880]: I0126 16:13:29.391632 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:29 crc kubenswrapper[4880]: I0126 16:13:29.392099 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:29 crc kubenswrapper[4880]: I0126 16:13:29.416633 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:29 crc kubenswrapper[4880]: I0126 16:13:29.577107 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:29 crc kubenswrapper[4880]: I0126 16:13:29.856616 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:31 crc kubenswrapper[4880]: I0126 16:13:31.562882 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-lrwwt" podUID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" containerName="registry-server" containerID="cri-o://aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414" gracePeriod=2 Jan 26 16:13:31 crc kubenswrapper[4880]: I0126 16:13:31.932193 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-68688768b9-p2q6k_cf21b426-5786-4406-be9a-469a13a1551b/operator/0.log" Jan 26 16:13:31 crc kubenswrapper[4880]: I0126 16:13:31.979697 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.143782 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2rqp\" (UniqueName: \"kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp\") pod \"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3\" (UID: \"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3\") " Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.149069 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp" (OuterVolumeSpecName: "kube-api-access-p2rqp") pod "e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" (UID: "e907e56d-9f3c-408d-b8bc-506d1d9dd7a3"). InnerVolumeSpecName "kube-api-access-p2rqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.245576 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2rqp\" (UniqueName: \"kubernetes.io/projected/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3-kube-api-access-p2rqp\") on node \"crc\" DevicePath \"\"" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.614047 4880 generic.go:334] "Generic (PLEG): container finished" podID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" containerID="aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414" exitCode=0 Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.614104 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-lrwwt" event={"ID":"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3","Type":"ContainerDied","Data":"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414"} Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.614137 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-lrwwt" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.614173 4880 scope.go:117] "RemoveContainer" containerID="aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.614145 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-lrwwt" event={"ID":"e907e56d-9f3c-408d-b8bc-506d1d9dd7a3","Type":"ContainerDied","Data":"e3477c1f9dcb915a4bad3bd6c8d8caf41e88b4a189c8e790d4acb322e779c248"} Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.638858 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.646169 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-lrwwt"] Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.648061 4880 scope.go:117] "RemoveContainer" containerID="aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414" Jan 26 16:13:32 crc kubenswrapper[4880]: E0126 16:13:32.648898 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414\": container with ID starting with aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414 not found: ID does not exist" containerID="aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414" Jan 26 16:13:32 crc kubenswrapper[4880]: I0126 16:13:32.649016 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414"} err="failed to get container status \"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414\": rpc error: code = NotFound desc = could not find container \"aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414\": container with ID starting with aef277c0d6ee74422b239f8c10f7d08a308d56c71b3c79f0b9ece523f6cca414 not found: ID does not exist" Jan 26 16:13:34 crc kubenswrapper[4880]: I0126 16:13:34.224935 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" path="/var/lib/kubelet/pods/e907e56d-9f3c-408d-b8bc-506d1d9dd7a3/volumes" Jan 26 16:13:35 crc kubenswrapper[4880]: I0126 16:13:35.747233 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-77c9d9f969-8jlvw_695785ce-59d7-4136-83c3-70ad2ab08f8b/operator/0.log" Jan 26 16:13:36 crc kubenswrapper[4880]: I0126 16:13:36.020364 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_23a071f2-a160-4871-bc14-cae91d0c08c9/qdr/0.log" Jan 26 16:13:41 crc kubenswrapper[4880]: I0126 16:13:41.099691 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:13:41 crc kubenswrapper[4880]: I0126 16:13:41.102014 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.162010 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7k6tt/must-gather-6k92m"] Jan 26 16:14:00 crc kubenswrapper[4880]: E0126 16:14:00.163130 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" containerName="registry-server" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.163160 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" containerName="registry-server" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.163403 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="e907e56d-9f3c-408d-b8bc-506d1d9dd7a3" containerName="registry-server" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.164794 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.167644 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7k6tt"/"openshift-service-ca.crt" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.178472 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7k6tt/must-gather-6k92m"] Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.183406 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7k6tt"/"kube-root-ca.crt" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.216235 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.216308 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfct7\" (UniqueName: \"kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.321780 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.321863 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfct7\" (UniqueName: \"kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.322619 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.347038 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfct7\" (UniqueName: \"kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7\") pod \"must-gather-6k92m\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.488608 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:14:00 crc kubenswrapper[4880]: I0126 16:14:00.925213 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7k6tt/must-gather-6k92m"] Jan 26 16:14:01 crc kubenswrapper[4880]: I0126 16:14:01.251426 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7k6tt/must-gather-6k92m" event={"ID":"0855de7d-2de7-423f-ae33-56a5d832d0da","Type":"ContainerStarted","Data":"c48a548780b3a931ae0c0b30dacda330fcfaf7c88094c6999a6c22e14218430c"} Jan 26 16:14:10 crc kubenswrapper[4880]: I0126 16:14:10.700963 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:14:10 crc kubenswrapper[4880]: I0126 16:14:10.701749 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:14:10 crc kubenswrapper[4880]: I0126 16:14:10.702070 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:14:10 crc kubenswrapper[4880]: I0126 16:14:10.703014 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:14:10 crc kubenswrapper[4880]: I0126 16:14:10.703128 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a" gracePeriod=600 Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.350128 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a" exitCode=0 Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.350325 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a"} Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.350490 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"6b7c708f5dd80ea2af55c819876bb426374a03da848f859d0c62e9f61c8596e1"} Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.350518 4880 scope.go:117] "RemoveContainer" containerID="4ae6a1c2569bb183c278e3306a62a1475002f7171aeedcca14880299354fdd56" Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.358758 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7k6tt/must-gather-6k92m" event={"ID":"0855de7d-2de7-423f-ae33-56a5d832d0da","Type":"ContainerStarted","Data":"4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200"} Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.358817 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7k6tt/must-gather-6k92m" event={"ID":"0855de7d-2de7-423f-ae33-56a5d832d0da","Type":"ContainerStarted","Data":"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e"} Jan 26 16:14:11 crc kubenswrapper[4880]: I0126 16:14:11.389589 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7k6tt/must-gather-6k92m" podStartSLOduration=1.9614661070000001 podStartE2EDuration="11.389556537s" podCreationTimestamp="2026-01-26 16:14:00 +0000 UTC" firstStartedPulling="2026-01-26 16:14:00.93757469 +0000 UTC m=+1166.423303397" lastFinishedPulling="2026-01-26 16:14:10.36566512 +0000 UTC m=+1175.851393827" observedRunningTime="2026-01-26 16:14:11.385893138 +0000 UTC m=+1176.871621845" watchObservedRunningTime="2026-01-26 16:14:11.389556537 +0000 UTC m=+1176.875285244" Jan 26 16:14:58 crc kubenswrapper[4880]: I0126 16:14:58.263744 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-54k89_a91a0129-b4bd-4acd-bceb-3091d5714089/control-plane-machine-set-operator/0.log" Jan 26 16:14:58 crc kubenswrapper[4880]: I0126 16:14:58.472746 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f9lnb_e96b5221-3d1f-4f21-bb0f-4b01373ffdb0/kube-rbac-proxy/0.log" Jan 26 16:14:58 crc kubenswrapper[4880]: I0126 16:14:58.473536 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f9lnb_e96b5221-3d1f-4f21-bb0f-4b01373ffdb0/machine-api-operator/0.log" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.153195 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n"] Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.154646 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.157550 4880 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.158012 4880 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.175038 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n"] Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.263860 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brr8h\" (UniqueName: \"kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.263948 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.264001 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.366769 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.366871 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brr8h\" (UniqueName: \"kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.366977 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.367890 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.378667 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.383833 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brr8h\" (UniqueName: \"kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h\") pod \"collect-profiles-29490735-4pf4n\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.489380 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.756854 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n"] Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.934394 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" event={"ID":"945bcfbb-28cf-426c-b221-e307a9897d64","Type":"ContainerStarted","Data":"1eec02b0894a6efca893f99acb231a2b5578791ea2a497392dcf8c3e42ad7579"} Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.934816 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" event={"ID":"945bcfbb-28cf-426c-b221-e307a9897d64","Type":"ContainerStarted","Data":"63b60da231377f0b4d7cd08ff591d0d18d2a37d3bab66486269e4ebaa88431ef"} Jan 26 16:15:00 crc kubenswrapper[4880]: I0126 16:15:00.955706 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" podStartSLOduration=0.955656218 podStartE2EDuration="955.656218ms" podCreationTimestamp="2026-01-26 16:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 16:15:00.948903314 +0000 UTC m=+1226.434632011" watchObservedRunningTime="2026-01-26 16:15:00.955656218 +0000 UTC m=+1226.441384925" Jan 26 16:15:01 crc kubenswrapper[4880]: I0126 16:15:01.943200 4880 generic.go:334] "Generic (PLEG): container finished" podID="945bcfbb-28cf-426c-b221-e307a9897d64" containerID="1eec02b0894a6efca893f99acb231a2b5578791ea2a497392dcf8c3e42ad7579" exitCode=0 Jan 26 16:15:01 crc kubenswrapper[4880]: I0126 16:15:01.943259 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" event={"ID":"945bcfbb-28cf-426c-b221-e307a9897d64","Type":"ContainerDied","Data":"1eec02b0894a6efca893f99acb231a2b5578791ea2a497392dcf8c3e42ad7579"} Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.263581 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.411888 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume\") pod \"945bcfbb-28cf-426c-b221-e307a9897d64\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.411950 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume\") pod \"945bcfbb-28cf-426c-b221-e307a9897d64\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.411978 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brr8h\" (UniqueName: \"kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h\") pod \"945bcfbb-28cf-426c-b221-e307a9897d64\" (UID: \"945bcfbb-28cf-426c-b221-e307a9897d64\") " Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.412887 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume" (OuterVolumeSpecName: "config-volume") pod "945bcfbb-28cf-426c-b221-e307a9897d64" (UID: "945bcfbb-28cf-426c-b221-e307a9897d64"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.423889 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "945bcfbb-28cf-426c-b221-e307a9897d64" (UID: "945bcfbb-28cf-426c-b221-e307a9897d64"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.423983 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h" (OuterVolumeSpecName: "kube-api-access-brr8h") pod "945bcfbb-28cf-426c-b221-e307a9897d64" (UID: "945bcfbb-28cf-426c-b221-e307a9897d64"). InnerVolumeSpecName "kube-api-access-brr8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.513371 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brr8h\" (UniqueName: \"kubernetes.io/projected/945bcfbb-28cf-426c-b221-e307a9897d64-kube-api-access-brr8h\") on node \"crc\" DevicePath \"\"" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.513427 4880 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/945bcfbb-28cf-426c-b221-e307a9897d64-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.513455 4880 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/945bcfbb-28cf-426c-b221-e307a9897d64-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.957992 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" event={"ID":"945bcfbb-28cf-426c-b221-e307a9897d64","Type":"ContainerDied","Data":"63b60da231377f0b4d7cd08ff591d0d18d2a37d3bab66486269e4ebaa88431ef"} Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.958046 4880 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63b60da231377f0b4d7cd08ff591d0d18d2a37d3bab66486269e4ebaa88431ef" Jan 26 16:15:03 crc kubenswrapper[4880]: I0126 16:15:03.958109 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490735-4pf4n" Jan 26 16:15:10 crc kubenswrapper[4880]: I0126 16:15:10.399741 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-fq6ff_e5326281-b231-4a2d-a1de-608c5815e977/cert-manager-controller/0.log" Jan 26 16:15:10 crc kubenswrapper[4880]: I0126 16:15:10.580502 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-v28tf_b957e4f4-7d0c-428c-bf37-4f4362d0e4bd/cert-manager-webhook/0.log" Jan 26 16:15:10 crc kubenswrapper[4880]: I0126 16:15:10.619405 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-snnps_94c919c2-b2ab-4bdd-bc26-dd8bb8f1c2ee/cert-manager-cainjector/0.log" Jan 26 16:15:23 crc kubenswrapper[4880]: I0126 16:15:23.796789 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-l4qc7_2bf4440d-678a-4d07-95aa-e7aed961c47c/prometheus-operator/0.log" Jan 26 16:15:23 crc kubenswrapper[4880]: I0126 16:15:23.903167 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-664d6f6879-b25tf_6d2ecd45-4798-48f7-a852-6007f00a1720/prometheus-operator-admission-webhook/0.log" Jan 26 16:15:23 crc kubenswrapper[4880]: I0126 16:15:23.994042 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm_48492eaa-8a65-43ee-8658-917a89a5be96/prometheus-operator-admission-webhook/0.log" Jan 26 16:15:24 crc kubenswrapper[4880]: I0126 16:15:24.081397 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-t4mkq_b3f672eb-d7d9-4316-a911-da9f76df698a/operator/0.log" Jan 26 16:15:24 crc kubenswrapper[4880]: I0126 16:15:24.206032 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-dwwh5_611b80dc-f8d4-4043-ae3b-974b03133bb3/perses-operator/0.log" Jan 26 16:15:37 crc kubenswrapper[4880]: I0126 16:15:37.777585 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/util/0.log" Jan 26 16:15:37 crc kubenswrapper[4880]: I0126 16:15:37.956367 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/util/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.008144 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.029560 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.186504 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/util/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.187329 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/extract/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.216932 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931azrs7k_4d6d78ab-29ae-4468-9a9d-9e44ec38278c/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.398554 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/util/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.570858 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.573910 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/util/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.581731 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.740143 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/pull/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.752870 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/util/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.753711 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fx8dgl_a039001a-3baa-48a4-9181-235e48958939/extract/0.log" Jan 26 16:15:38 crc kubenswrapper[4880]: I0126 16:15:38.884490 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/util/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.190466 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/util/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.195733 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/pull/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.219395 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/pull/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.502152 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/util/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.518207 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/pull/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.523560 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5epz5hr_3d4820bd-1290-4cea-90fb-7ed5b43d5d55/extract/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.656014 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/util/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.839220 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/util/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.848965 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/pull/0.log" Jan 26 16:15:39 crc kubenswrapper[4880]: I0126 16:15:39.862854 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/pull/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.035611 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/util/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.038313 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/pull/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.069647 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08kjxm2_7bcbfaec-4cc2-425e-be37-1556c5f92fdb/extract/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.194099 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-utilities/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.873361 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-content/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.885461 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-content/0.log" Jan 26 16:15:40 crc kubenswrapper[4880]: I0126 16:15:40.904889 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-utilities/0.log" Jan 26 16:15:41 crc kubenswrapper[4880]: I0126 16:15:41.660850 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-utilities/0.log" Jan 26 16:15:41 crc kubenswrapper[4880]: I0126 16:15:41.725116 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/extract-content/0.log" Jan 26 16:15:41 crc kubenswrapper[4880]: I0126 16:15:41.863487 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-utilities/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.059250 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p7xk2_6aa3e1ef-b403-4bc1-a5b7-87dbe0490bac/registry-server/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.151005 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-utilities/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.157919 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-content/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.199039 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-content/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.392045 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-content/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.395933 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/extract-utilities/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.474544 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-zlnn7_c376a5ed-f9eb-4592-b6b5-2e021c59d978/marketplace-operator/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.725472 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zksbn_4176b049-00e6-4f38-8045-4d84d2926272/registry-server/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.755047 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-utilities/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.844389 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-content/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.845752 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-utilities/0.log" Jan 26 16:15:42 crc kubenswrapper[4880]: I0126 16:15:42.888830 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-content/0.log" Jan 26 16:15:43 crc kubenswrapper[4880]: I0126 16:15:43.022010 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-utilities/0.log" Jan 26 16:15:43 crc kubenswrapper[4880]: I0126 16:15:43.047659 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/extract-content/0.log" Jan 26 16:15:43 crc kubenswrapper[4880]: I0126 16:15:43.375670 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mkk48_9ec96499-4381-4628-9c64-1a8de843bf6d/registry-server/0.log" Jan 26 16:15:56 crc kubenswrapper[4880]: I0126 16:15:56.160674 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-664d6f6879-b25tf_6d2ecd45-4798-48f7-a852-6007f00a1720/prometheus-operator-admission-webhook/0.log" Jan 26 16:15:56 crc kubenswrapper[4880]: I0126 16:15:56.222779 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-664d6f6879-sqbnm_48492eaa-8a65-43ee-8658-917a89a5be96/prometheus-operator-admission-webhook/0.log" Jan 26 16:15:56 crc kubenswrapper[4880]: I0126 16:15:56.230332 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-l4qc7_2bf4440d-678a-4d07-95aa-e7aed961c47c/prometheus-operator/0.log" Jan 26 16:15:56 crc kubenswrapper[4880]: I0126 16:15:56.331159 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-t4mkq_b3f672eb-d7d9-4316-a911-da9f76df698a/operator/0.log" Jan 26 16:15:56 crc kubenswrapper[4880]: I0126 16:15:56.391303 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-dwwh5_611b80dc-f8d4-4043-ae3b-974b03133bb3/perses-operator/0.log" Jan 26 16:16:40 crc kubenswrapper[4880]: I0126 16:16:40.700636 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:16:40 crc kubenswrapper[4880]: I0126 16:16:40.701258 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:16:54 crc kubenswrapper[4880]: I0126 16:16:54.702173 4880 generic.go:334] "Generic (PLEG): container finished" podID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerID="800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e" exitCode=0 Jan 26 16:16:54 crc kubenswrapper[4880]: I0126 16:16:54.702250 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7k6tt/must-gather-6k92m" event={"ID":"0855de7d-2de7-423f-ae33-56a5d832d0da","Type":"ContainerDied","Data":"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e"} Jan 26 16:16:54 crc kubenswrapper[4880]: I0126 16:16:54.703475 4880 scope.go:117] "RemoveContainer" containerID="800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e" Jan 26 16:16:55 crc kubenswrapper[4880]: I0126 16:16:55.197825 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7k6tt_must-gather-6k92m_0855de7d-2de7-423f-ae33-56a5d832d0da/gather/0.log" Jan 26 16:17:02 crc kubenswrapper[4880]: I0126 16:17:02.657536 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7k6tt/must-gather-6k92m"] Jan 26 16:17:02 crc kubenswrapper[4880]: I0126 16:17:02.658401 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7k6tt/must-gather-6k92m" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="copy" containerID="cri-o://4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200" gracePeriod=2 Jan 26 16:17:02 crc kubenswrapper[4880]: I0126 16:17:02.665546 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7k6tt/must-gather-6k92m"] Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.834015 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7k6tt_must-gather-6k92m_0855de7d-2de7-423f-ae33-56a5d832d0da/copy/0.log" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.834914 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.873708 4880 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7k6tt_must-gather-6k92m_0855de7d-2de7-423f-ae33-56a5d832d0da/copy/0.log" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.874263 4880 generic.go:334] "Generic (PLEG): container finished" podID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerID="4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200" exitCode=143 Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.874343 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7k6tt/must-gather-6k92m" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.874358 4880 scope.go:117] "RemoveContainer" containerID="4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.896471 4880 scope.go:117] "RemoveContainer" containerID="800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.912287 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output\") pod \"0855de7d-2de7-423f-ae33-56a5d832d0da\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.912380 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfct7\" (UniqueName: \"kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7\") pod \"0855de7d-2de7-423f-ae33-56a5d832d0da\" (UID: \"0855de7d-2de7-423f-ae33-56a5d832d0da\") " Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.920353 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7" (OuterVolumeSpecName: "kube-api-access-bfct7") pod "0855de7d-2de7-423f-ae33-56a5d832d0da" (UID: "0855de7d-2de7-423f-ae33-56a5d832d0da"). InnerVolumeSpecName "kube-api-access-bfct7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.956353 4880 scope.go:117] "RemoveContainer" containerID="4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200" Jan 26 16:17:03 crc kubenswrapper[4880]: E0126 16:17:03.956951 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200\": container with ID starting with 4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200 not found: ID does not exist" containerID="4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.956991 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200"} err="failed to get container status \"4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200\": rpc error: code = NotFound desc = could not find container \"4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200\": container with ID starting with 4327a27c729534b69f8d477a27a7f3cdf9f37c8230a776f8d925e5aac391e200 not found: ID does not exist" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.957013 4880 scope.go:117] "RemoveContainer" containerID="800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e" Jan 26 16:17:03 crc kubenswrapper[4880]: E0126 16:17:03.958240 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e\": container with ID starting with 800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e not found: ID does not exist" containerID="800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.958264 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e"} err="failed to get container status \"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e\": rpc error: code = NotFound desc = could not find container \"800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e\": container with ID starting with 800f0b7ce897747154313d964e192047aeb7392a84d19fe412cad63e3516fa9e not found: ID does not exist" Jan 26 16:17:03 crc kubenswrapper[4880]: I0126 16:17:03.978823 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0855de7d-2de7-423f-ae33-56a5d832d0da" (UID: "0855de7d-2de7-423f-ae33-56a5d832d0da"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:17:04 crc kubenswrapper[4880]: I0126 16:17:04.013513 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfct7\" (UniqueName: \"kubernetes.io/projected/0855de7d-2de7-423f-ae33-56a5d832d0da-kube-api-access-bfct7\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:04 crc kubenswrapper[4880]: I0126 16:17:04.013559 4880 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0855de7d-2de7-423f-ae33-56a5d832d0da-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:04 crc kubenswrapper[4880]: I0126 16:17:04.229641 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" path="/var/lib/kubelet/pods/0855de7d-2de7-423f-ae33-56a5d832d0da/volumes" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.168942 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:05 crc kubenswrapper[4880]: E0126 16:17:05.170749 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="copy" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.170869 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="copy" Jan 26 16:17:05 crc kubenswrapper[4880]: E0126 16:17:05.170943 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="945bcfbb-28cf-426c-b221-e307a9897d64" containerName="collect-profiles" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.171011 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="945bcfbb-28cf-426c-b221-e307a9897d64" containerName="collect-profiles" Jan 26 16:17:05 crc kubenswrapper[4880]: E0126 16:17:05.171088 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="gather" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.171146 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="gather" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.171367 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="945bcfbb-28cf-426c-b221-e307a9897d64" containerName="collect-profiles" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.171482 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="copy" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.171572 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="0855de7d-2de7-423f-ae33-56a5d832d0da" containerName="gather" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.172625 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.188034 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.410624 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.411402 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnzk8\" (UniqueName: \"kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.411617 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.512936 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnzk8\" (UniqueName: \"kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.513057 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.513126 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.513760 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.513817 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.541034 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnzk8\" (UniqueName: \"kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8\") pod \"redhat-operators-z7xkq\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:05 crc kubenswrapper[4880]: I0126 16:17:05.793629 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:06 crc kubenswrapper[4880]: I0126 16:17:06.049227 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:06 crc kubenswrapper[4880]: I0126 16:17:06.898803 4880 generic.go:334] "Generic (PLEG): container finished" podID="433035be-d2aa-45b8-b243-bace8b49f86d" containerID="1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87" exitCode=0 Jan 26 16:17:06 crc kubenswrapper[4880]: I0126 16:17:06.899017 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerDied","Data":"1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87"} Jan 26 16:17:06 crc kubenswrapper[4880]: I0126 16:17:06.899156 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerStarted","Data":"5cdb943171f06db5a00f9ee939f6eff9e6865785c6658f8622e69d52a2ebb5ba"} Jan 26 16:17:06 crc kubenswrapper[4880]: I0126 16:17:06.901006 4880 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 16:17:08 crc kubenswrapper[4880]: I0126 16:17:08.929116 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerStarted","Data":"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb"} Jan 26 16:17:09 crc kubenswrapper[4880]: I0126 16:17:09.941416 4880 generic.go:334] "Generic (PLEG): container finished" podID="433035be-d2aa-45b8-b243-bace8b49f86d" containerID="5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb" exitCode=0 Jan 26 16:17:09 crc kubenswrapper[4880]: I0126 16:17:09.941558 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerDied","Data":"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb"} Jan 26 16:17:10 crc kubenswrapper[4880]: I0126 16:17:10.700200 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:17:10 crc kubenswrapper[4880]: I0126 16:17:10.700520 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:17:10 crc kubenswrapper[4880]: I0126 16:17:10.957307 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerStarted","Data":"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1"} Jan 26 16:17:10 crc kubenswrapper[4880]: I0126 16:17:10.985363 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z7xkq" podStartSLOduration=2.468179606 podStartE2EDuration="5.985289807s" podCreationTimestamp="2026-01-26 16:17:05 +0000 UTC" firstStartedPulling="2026-01-26 16:17:06.900623205 +0000 UTC m=+1352.386351912" lastFinishedPulling="2026-01-26 16:17:10.417733406 +0000 UTC m=+1355.903462113" observedRunningTime="2026-01-26 16:17:10.980804238 +0000 UTC m=+1356.466532945" watchObservedRunningTime="2026-01-26 16:17:10.985289807 +0000 UTC m=+1356.471018554" Jan 26 16:17:15 crc kubenswrapper[4880]: I0126 16:17:15.795659 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:15 crc kubenswrapper[4880]: I0126 16:17:15.796036 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.737180 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.739462 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.752232 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.838814 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.839019 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.839100 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh6gc\" (UniqueName: \"kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.839139 4880 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z7xkq" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="registry-server" probeResult="failure" output=< Jan 26 16:17:16 crc kubenswrapper[4880]: timeout: failed to connect service ":50051" within 1s Jan 26 16:17:16 crc kubenswrapper[4880]: > Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.940133 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.940231 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.940260 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh6gc\" (UniqueName: \"kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.940810 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.940924 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:16 crc kubenswrapper[4880]: I0126 16:17:16.963642 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh6gc\" (UniqueName: \"kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc\") pod \"certified-operators-8v4ps\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:17 crc kubenswrapper[4880]: I0126 16:17:17.066335 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:17 crc kubenswrapper[4880]: I0126 16:17:17.357750 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:18 crc kubenswrapper[4880]: I0126 16:17:18.010659 4880 generic.go:334] "Generic (PLEG): container finished" podID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerID="33fc00acc84f3563a85df7f5f152b0918d567a979db2775dd8b3af3dad2d3ecf" exitCode=0 Jan 26 16:17:18 crc kubenswrapper[4880]: I0126 16:17:18.010805 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerDied","Data":"33fc00acc84f3563a85df7f5f152b0918d567a979db2775dd8b3af3dad2d3ecf"} Jan 26 16:17:18 crc kubenswrapper[4880]: I0126 16:17:18.011008 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerStarted","Data":"341629b006e086daa4fd799f895361e38991cb96cfc2d73cb3203edd99e7ad34"} Jan 26 16:17:19 crc kubenswrapper[4880]: I0126 16:17:19.020509 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerStarted","Data":"057f0574ec76d570db0229f37ade0db5bd802b1c8e2d83ef0f54f9d2edc4cdeb"} Jan 26 16:17:21 crc kubenswrapper[4880]: I0126 16:17:21.040106 4880 generic.go:334] "Generic (PLEG): container finished" podID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerID="057f0574ec76d570db0229f37ade0db5bd802b1c8e2d83ef0f54f9d2edc4cdeb" exitCode=0 Jan 26 16:17:21 crc kubenswrapper[4880]: I0126 16:17:21.041256 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerDied","Data":"057f0574ec76d570db0229f37ade0db5bd802b1c8e2d83ef0f54f9d2edc4cdeb"} Jan 26 16:17:22 crc kubenswrapper[4880]: I0126 16:17:22.049488 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerStarted","Data":"1871d69e7873d43d2b09a149691b1b248101aa3f05008e01c355d9f32160bd04"} Jan 26 16:17:22 crc kubenswrapper[4880]: I0126 16:17:22.073083 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8v4ps" podStartSLOduration=2.5443273680000003 podStartE2EDuration="6.073062509s" podCreationTimestamp="2026-01-26 16:17:16 +0000 UTC" firstStartedPulling="2026-01-26 16:17:18.012731205 +0000 UTC m=+1363.498459922" lastFinishedPulling="2026-01-26 16:17:21.541466356 +0000 UTC m=+1367.027195063" observedRunningTime="2026-01-26 16:17:22.067532035 +0000 UTC m=+1367.553260742" watchObservedRunningTime="2026-01-26 16:17:22.073062509 +0000 UTC m=+1367.558791216" Jan 26 16:17:25 crc kubenswrapper[4880]: I0126 16:17:25.839019 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:25 crc kubenswrapper[4880]: I0126 16:17:25.882173 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:26 crc kubenswrapper[4880]: I0126 16:17:26.076400 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.066700 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.066798 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.085132 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z7xkq" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="registry-server" containerID="cri-o://77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1" gracePeriod=2 Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.121801 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.198158 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:27 crc kubenswrapper[4880]: I0126 16:17:27.991747 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.095895 4880 generic.go:334] "Generic (PLEG): container finished" podID="433035be-d2aa-45b8-b243-bace8b49f86d" containerID="77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1" exitCode=0 Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.095967 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7xkq" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.096015 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerDied","Data":"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1"} Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.096055 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7xkq" event={"ID":"433035be-d2aa-45b8-b243-bace8b49f86d","Type":"ContainerDied","Data":"5cdb943171f06db5a00f9ee939f6eff9e6865785c6658f8622e69d52a2ebb5ba"} Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.096084 4880 scope.go:117] "RemoveContainer" containerID="77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.105314 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities\") pod \"433035be-d2aa-45b8-b243-bace8b49f86d\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.105746 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnzk8\" (UniqueName: \"kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8\") pod \"433035be-d2aa-45b8-b243-bace8b49f86d\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.105814 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content\") pod \"433035be-d2aa-45b8-b243-bace8b49f86d\" (UID: \"433035be-d2aa-45b8-b243-bace8b49f86d\") " Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.106378 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities" (OuterVolumeSpecName: "utilities") pod "433035be-d2aa-45b8-b243-bace8b49f86d" (UID: "433035be-d2aa-45b8-b243-bace8b49f86d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.113739 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8" (OuterVolumeSpecName: "kube-api-access-pnzk8") pod "433035be-d2aa-45b8-b243-bace8b49f86d" (UID: "433035be-d2aa-45b8-b243-bace8b49f86d"). InnerVolumeSpecName "kube-api-access-pnzk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.117891 4880 scope.go:117] "RemoveContainer" containerID="5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.171397 4880 scope.go:117] "RemoveContainer" containerID="1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.189731 4880 scope.go:117] "RemoveContainer" containerID="77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1" Jan 26 16:17:28 crc kubenswrapper[4880]: E0126 16:17:28.190589 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1\": container with ID starting with 77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1 not found: ID does not exist" containerID="77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.190716 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1"} err="failed to get container status \"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1\": rpc error: code = NotFound desc = could not find container \"77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1\": container with ID starting with 77aaa9a033356290c329df39e09df30e4434c821b4a1471b45a4ca61f46fbda1 not found: ID does not exist" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.190810 4880 scope.go:117] "RemoveContainer" containerID="5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb" Jan 26 16:17:28 crc kubenswrapper[4880]: E0126 16:17:28.191291 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb\": container with ID starting with 5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb not found: ID does not exist" containerID="5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.191388 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb"} err="failed to get container status \"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb\": rpc error: code = NotFound desc = could not find container \"5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb\": container with ID starting with 5d15032f006f86dc4caf0e68280c7bf5375b7b525095f90da6a5ee0f13d0b5bb not found: ID does not exist" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.191493 4880 scope.go:117] "RemoveContainer" containerID="1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87" Jan 26 16:17:28 crc kubenswrapper[4880]: E0126 16:17:28.191870 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87\": container with ID starting with 1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87 not found: ID does not exist" containerID="1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.191933 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87"} err="failed to get container status \"1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87\": rpc error: code = NotFound desc = could not find container \"1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87\": container with ID starting with 1941ec48869379d4a4aeeb3ab9f3d99d4972682554c0ca2be50236f7839ccf87 not found: ID does not exist" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.207464 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnzk8\" (UniqueName: \"kubernetes.io/projected/433035be-d2aa-45b8-b243-bace8b49f86d-kube-api-access-pnzk8\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.207511 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.229343 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "433035be-d2aa-45b8-b243-bace8b49f86d" (UID: "433035be-d2aa-45b8-b243-bace8b49f86d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.308334 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/433035be-d2aa-45b8-b243-bace8b49f86d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.417703 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.430936 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z7xkq"] Jan 26 16:17:28 crc kubenswrapper[4880]: I0126 16:17:28.874965 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:29 crc kubenswrapper[4880]: I0126 16:17:29.104335 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8v4ps" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="registry-server" containerID="cri-o://1871d69e7873d43d2b09a149691b1b248101aa3f05008e01c355d9f32160bd04" gracePeriod=2 Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.113682 4880 generic.go:334] "Generic (PLEG): container finished" podID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerID="1871d69e7873d43d2b09a149691b1b248101aa3f05008e01c355d9f32160bd04" exitCode=0 Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.113797 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerDied","Data":"1871d69e7873d43d2b09a149691b1b248101aa3f05008e01c355d9f32160bd04"} Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.223463 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" path="/var/lib/kubelet/pods/433035be-d2aa-45b8-b243-bace8b49f86d/volumes" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.707261 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.855421 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content\") pod \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.855603 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities\") pod \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.855659 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh6gc\" (UniqueName: \"kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc\") pod \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\" (UID: \"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3\") " Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.856547 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities" (OuterVolumeSpecName: "utilities") pod "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" (UID: "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.861631 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc" (OuterVolumeSpecName: "kube-api-access-kh6gc") pod "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" (UID: "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3"). InnerVolumeSpecName "kube-api-access-kh6gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.908646 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" (UID: "08b04b77-d44e-4d28-b0d7-cc4c8743b0f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.957261 4880 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.957623 4880 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:30 crc kubenswrapper[4880]: I0126 16:17:30.957708 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh6gc\" (UniqueName: \"kubernetes.io/projected/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3-kube-api-access-kh6gc\") on node \"crc\" DevicePath \"\"" Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.122773 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8v4ps" event={"ID":"08b04b77-d44e-4d28-b0d7-cc4c8743b0f3","Type":"ContainerDied","Data":"341629b006e086daa4fd799f895361e38991cb96cfc2d73cb3203edd99e7ad34"} Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.122844 4880 scope.go:117] "RemoveContainer" containerID="1871d69e7873d43d2b09a149691b1b248101aa3f05008e01c355d9f32160bd04" Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.122840 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8v4ps" Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.148878 4880 scope.go:117] "RemoveContainer" containerID="057f0574ec76d570db0229f37ade0db5bd802b1c8e2d83ef0f54f9d2edc4cdeb" Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.158890 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.164385 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8v4ps"] Jan 26 16:17:31 crc kubenswrapper[4880]: I0126 16:17:31.188020 4880 scope.go:117] "RemoveContainer" containerID="33fc00acc84f3563a85df7f5f152b0918d567a979db2775dd8b3af3dad2d3ecf" Jan 26 16:17:32 crc kubenswrapper[4880]: I0126 16:17:32.223524 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" path="/var/lib/kubelet/pods/08b04b77-d44e-4d28-b0d7-cc4c8743b0f3/volumes" Jan 26 16:17:40 crc kubenswrapper[4880]: I0126 16:17:40.708970 4880 patch_prober.go:28] interesting pod/machine-config-daemon-glj8k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 16:17:40 crc kubenswrapper[4880]: I0126 16:17:40.709656 4880 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 16:17:40 crc kubenswrapper[4880]: I0126 16:17:40.709721 4880 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" Jan 26 16:17:40 crc kubenswrapper[4880]: I0126 16:17:40.710602 4880 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b7c708f5dd80ea2af55c819876bb426374a03da848f859d0c62e9f61c8596e1"} pod="openshift-machine-config-operator/machine-config-daemon-glj8k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 16:17:40 crc kubenswrapper[4880]: I0126 16:17:40.710654 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" podUID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerName="machine-config-daemon" containerID="cri-o://6b7c708f5dd80ea2af55c819876bb426374a03da848f859d0c62e9f61c8596e1" gracePeriod=600 Jan 26 16:17:41 crc kubenswrapper[4880]: I0126 16:17:41.203564 4880 generic.go:334] "Generic (PLEG): container finished" podID="db202b43-2be9-4b6c-87b7-b2c5ef2df984" containerID="6b7c708f5dd80ea2af55c819876bb426374a03da848f859d0c62e9f61c8596e1" exitCode=0 Jan 26 16:17:41 crc kubenswrapper[4880]: I0126 16:17:41.203657 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerDied","Data":"6b7c708f5dd80ea2af55c819876bb426374a03da848f859d0c62e9f61c8596e1"} Jan 26 16:17:41 crc kubenswrapper[4880]: I0126 16:17:41.204011 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-glj8k" event={"ID":"db202b43-2be9-4b6c-87b7-b2c5ef2df984","Type":"ContainerStarted","Data":"3250b231034f4465624843fb717697503e41f0f48e066c81a2371509bf982c96"} Jan 26 16:17:41 crc kubenswrapper[4880]: I0126 16:17:41.204041 4880 scope.go:117] "RemoveContainer" containerID="6249315386a4893ba86d098f4ede237fb65c7d55d3a8e26011d945ac37b88f3a" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.681592 4880 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.682836 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="extract-content" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.682875 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="extract-content" Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.682907 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="extract-utilities" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.682915 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="extract-utilities" Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.682960 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.682971 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.682985 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.682999 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.683011 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="extract-utilities" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.683021 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="extract-utilities" Jan 26 16:18:19 crc kubenswrapper[4880]: E0126 16:18:19.683041 4880 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="extract-content" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.683049 4880 state_mem.go:107] "Deleted CPUSet assignment" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="extract-content" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.683400 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="433035be-d2aa-45b8-b243-bace8b49f86d" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.683420 4880 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b04b77-d44e-4d28-b0d7-cc4c8743b0f3" containerName="registry-server" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.686083 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.706409 4880 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhqq\" (UniqueName: \"kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq\") pod \"infrawatch-operators-zdtv4\" (UID: \"29483343-d422-4f42-8722-1b8fc8f5b2d0\") " pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.712926 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:19 crc kubenswrapper[4880]: I0126 16:18:19.808810 4880 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhqq\" (UniqueName: \"kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq\") pod \"infrawatch-operators-zdtv4\" (UID: \"29483343-d422-4f42-8722-1b8fc8f5b2d0\") " pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:20 crc kubenswrapper[4880]: I0126 16:18:20.003879 4880 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhqq\" (UniqueName: \"kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq\") pod \"infrawatch-operators-zdtv4\" (UID: \"29483343-d422-4f42-8722-1b8fc8f5b2d0\") " pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:20 crc kubenswrapper[4880]: I0126 16:18:20.016610 4880 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:20 crc kubenswrapper[4880]: I0126 16:18:20.277189 4880 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:20 crc kubenswrapper[4880]: I0126 16:18:20.481162 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-zdtv4" event={"ID":"29483343-d422-4f42-8722-1b8fc8f5b2d0","Type":"ContainerStarted","Data":"9ac6d91895fa47fa1932e563545c02fb847c34dcc69b71568a4b4a8e51e23f80"} Jan 26 16:18:22 crc kubenswrapper[4880]: I0126 16:18:22.500238 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-zdtv4" event={"ID":"29483343-d422-4f42-8722-1b8fc8f5b2d0","Type":"ContainerStarted","Data":"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425"} Jan 26 16:18:22 crc kubenswrapper[4880]: I0126 16:18:22.536709 4880 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-zdtv4" podStartSLOduration=2.683836566 podStartE2EDuration="3.536675179s" podCreationTimestamp="2026-01-26 16:18:19 +0000 UTC" firstStartedPulling="2026-01-26 16:18:20.28491623 +0000 UTC m=+1425.770644937" lastFinishedPulling="2026-01-26 16:18:21.137754843 +0000 UTC m=+1426.623483550" observedRunningTime="2026-01-26 16:18:22.534079788 +0000 UTC m=+1428.019808495" watchObservedRunningTime="2026-01-26 16:18:22.536675179 +0000 UTC m=+1428.022403886" Jan 26 16:18:30 crc kubenswrapper[4880]: I0126 16:18:30.017867 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:30 crc kubenswrapper[4880]: I0126 16:18:30.018622 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:30 crc kubenswrapper[4880]: I0126 16:18:30.054054 4880 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:30 crc kubenswrapper[4880]: I0126 16:18:30.631682 4880 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:30 crc kubenswrapper[4880]: I0126 16:18:30.678174 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:32 crc kubenswrapper[4880]: I0126 16:18:32.616126 4880 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-zdtv4" podUID="29483343-d422-4f42-8722-1b8fc8f5b2d0" containerName="registry-server" containerID="cri-o://0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425" gracePeriod=2 Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.527982 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.624246 4880 generic.go:334] "Generic (PLEG): container finished" podID="29483343-d422-4f42-8722-1b8fc8f5b2d0" containerID="0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425" exitCode=0 Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.624330 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-zdtv4" event={"ID":"29483343-d422-4f42-8722-1b8fc8f5b2d0","Type":"ContainerDied","Data":"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425"} Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.624455 4880 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-zdtv4" event={"ID":"29483343-d422-4f42-8722-1b8fc8f5b2d0","Type":"ContainerDied","Data":"9ac6d91895fa47fa1932e563545c02fb847c34dcc69b71568a4b4a8e51e23f80"} Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.624502 4880 scope.go:117] "RemoveContainer" containerID="0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.624367 4880 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-zdtv4" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.632958 4880 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdhqq\" (UniqueName: \"kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq\") pod \"29483343-d422-4f42-8722-1b8fc8f5b2d0\" (UID: \"29483343-d422-4f42-8722-1b8fc8f5b2d0\") " Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.639720 4880 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq" (OuterVolumeSpecName: "kube-api-access-pdhqq") pod "29483343-d422-4f42-8722-1b8fc8f5b2d0" (UID: "29483343-d422-4f42-8722-1b8fc8f5b2d0"). InnerVolumeSpecName "kube-api-access-pdhqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.649032 4880 scope.go:117] "RemoveContainer" containerID="0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425" Jan 26 16:18:33 crc kubenswrapper[4880]: E0126 16:18:33.649633 4880 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425\": container with ID starting with 0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425 not found: ID does not exist" containerID="0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.649672 4880 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425"} err="failed to get container status \"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425\": rpc error: code = NotFound desc = could not find container \"0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425\": container with ID starting with 0269d5c289d524d458607ee5e1eec6c2de3c84bcc4607156fbcd6f8f311ab425 not found: ID does not exist" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.735088 4880 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdhqq\" (UniqueName: \"kubernetes.io/projected/29483343-d422-4f42-8722-1b8fc8f5b2d0-kube-api-access-pdhqq\") on node \"crc\" DevicePath \"\"" Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.954218 4880 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:33 crc kubenswrapper[4880]: I0126 16:18:33.963185 4880 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-zdtv4"] Jan 26 16:18:34 crc kubenswrapper[4880]: I0126 16:18:34.223294 4880 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29483343-d422-4f42-8722-1b8fc8f5b2d0" path="/var/lib/kubelet/pods/29483343-d422-4f42-8722-1b8fc8f5b2d0/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515135712030024442 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015135712031017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015135706544016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015135706544015467 5ustar corecore